US20070280550A1 - Lens defect correction - Google Patents
Lens defect correction Download PDFInfo
- Publication number
- US20070280550A1 US20070280550A1 US11/804,314 US80431407A US2007280550A1 US 20070280550 A1 US20070280550 A1 US 20070280550A1 US 80431407 A US80431407 A US 80431407A US 2007280550 A1 US2007280550 A1 US 2007280550A1
- Authority
- US
- United States
- Prior art keywords
- image
- lens
- primary
- focus
- electrical circuitry
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 230000007547 defect Effects 0.000 title description 12
- 238000012937 correction Methods 0.000 title description 3
- 239000002131 composite material Substances 0.000 claims abstract description 22
- 238000000034 method Methods 0.000 claims description 69
- 230000004044 response Effects 0.000 claims description 20
- 230000003287 optical effect Effects 0.000 claims description 15
- 230000005540 biological transmission Effects 0.000 claims description 12
- 238000004891 communication Methods 0.000 claims description 11
- 238000004590 computer program Methods 0.000 claims description 10
- 230000033001 locomotion Effects 0.000 claims description 10
- 230000004075 alteration Effects 0.000 claims description 9
- 238000010276 construction Methods 0.000 claims description 7
- 238000006073 displacement reaction Methods 0.000 claims description 4
- 238000013507 mapping Methods 0.000 claims description 3
- 238000013475 authorization Methods 0.000 claims 1
- 230000008569 process Effects 0.000 description 21
- 238000003384 imaging method Methods 0.000 description 14
- 238000012545 processing Methods 0.000 description 8
- 238000001514 detection method Methods 0.000 description 5
- 230000006870 function Effects 0.000 description 4
- 238000004519 manufacturing process Methods 0.000 description 4
- 238000010586 diagram Methods 0.000 description 3
- 230000007704 transition Effects 0.000 description 3
- 238000004458 analytical method Methods 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000003491 array Methods 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000007796 conventional method Methods 0.000 description 1
- 239000013078 crystal Substances 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000002156 mixing Methods 0.000 description 1
- 239000003607 modifier Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/67—Focus control based on electronic image sensor signals
- H04N23/673—Focus control based on electronic image sensor signals based on contrast or high frequency components of image signals, e.g. hill climbing method
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/60—Noise processing, e.g. detecting, correcting, reducing or removing noise
- H04N25/61—Noise processing, e.g. detecting, correcting, reducing or removing noise the noise originating only from the lens unit, e.g. flare, shading, vignetting or "cos4"
Definitions
- the present application relates, in general, to imaging.
- a method includes but is not limited to: capturing a primary image with a lens at a primary position, the lens having at least one deviation that exceeds a first tolerance from a target optical property; capturing another image with the lens at another position; determining at least one out-of-focus region of the primary image; determining a focus of at least one region of the other image relative to a focus of the at least one out-of-focus region of the primary image; and constructing a composite image in response to the at least one region of the other image having a sharper focus relative to the focus of the at least one out-of-focus region of the primary image.
- related systems include but are not limited to electro-mechanical systems (e.g., motors, actuators, circuitry, and/or programming) for effecting the herein-referenced method embodiments); the electrical circuitry can be virtually any combination of hardware, software, and/or firmware configured to effect the foregoing- referenced method embodiments depending upon the design choices of the system designer.
- electro-mechanical systems e.g., motors, actuators, circuitry, and/or programming
- a system includes but is not limited to: a photo-detector array; a lens having at least one deviation that exceeds a first tolerance from a target optical property; a controller configured to position said lens at a primary and another position relative to said photo-detector array and to cause an image capture signal at the primary and the other position; and an image construction unit configured to construct at least one out-of-focus region of a first image captured at the primary position with a more in-focus region of another image captured at the other position.
- FIG. 1 shows a front-plan view of image 100 of a person (e.g., person 202 of FIG. 2 ) projected onto photo-detector array 102 .
- FIG. 2 depicts a side-plan view of lens system 200 that can give rise to image 100 of FIG. 1 .
- FIG. 3 depicts a high level logic flowchart of a process.
- FIG. 4 depicts a side-plan view of the system of FIG. 2 wherein lens 204 has been moved in accordance with aspects of the process shown and described in relation to FIG. 3 .
- FIG. 5 illustrates another side-plan view of the system of FIG. 2 wherein lens 204 has been moved in accordance with aspects of the process shown and described in relation to FIG. 3 .
- FIG. 1 shown is a front-plan view of image 100 of a person (e.g., person 202 of FIG. 2 ) projected onto photo-detector array 102 .
- Image 100 is shown as distorted due to defects in a lens through which image 100 has been projected (e.g., lens 204 of lens system 200 of FIG. 2 ).
- First portion 104 of image 100 is illustrated as large and blurry, which can occur when a lens defect causes portion 104 of image 100 to come to a focus in front of a surface of photo-detector array 102 .
- third, and fourth portions 106 are illustrated as right sized, which can occur when the lens causes portions 106 of image 100 to correctly focus on an imaging surface of photo-detector array 102 .
- Fifth portion 108 is shown as small and faint, which can occur when a lens defect causes portion 108 of image 100 to come to a focus (virtual) behind an imaging surface of photo-detector array 102 .
- various lens defects could also cause the image to be distorted in x-y; those having skill in the art will also appreciate that different colored wavelengths of light can in and of themselves focus at different positions due to differences in refraction of the different colored wavelengths of light.
- FIG. 2 depicted is a side-plan view of lens system 200 that can give rise to image 100 of FIG. 1 .
- Lens 204 of lens system 200 is illustrated as located at a primary position and having defects that give rise to the five different portions of image 100 shown and described in relation to FIG. 1 .
- First portion 104 of image 100 is illustrated as focused in front of an imaging surface of photo-detector array 102 .
- Second, third, and fourth portions 106 are illustrated as right sized and focused on an imaging surface of photo-detector array 102 .
- Fifth portion 108 is shown as small and faint, and virtually focused behind an imaging surface of photo-detector array 102 .
- controller 208 is depicted as controlling the position of lens 204 of lens system 200 (e.g., via use of a feedback control subsystem).
- Image capture unit 206 is illustrated as receiving image data from photo-detector 102 and receiving control signals from controller 208 .
- Image capture unit 206 is shown as transmitting captured image information to focus detection unit 210 .
- Focus detection unit 210 is depicted as transmitting focus data to image construction unit 212 .
- Image construction unit 212 is illustrated as transmitting a composite image to image store/display unit 214 .
- Method step 300 shows the start of the process.
- Method step 302 depicts capturing a primary image with a lens at a primary position, the lens having at least one deviation that exceeds a first tolerance from a target optical property.
- the lens having at least one deviation that exceeds a first tolerance from a target optical property would be where the lens has at least one focal length that exceeds a first tolerance from a defined focal length (e.g., a defect that would produce portion 108 of image 100 at some place behind an imaging surface of photo-detector 102 or a defect that would produce portion 104 at some place in front of the imaging surface of photo-detector array 102 where the distance in front or back of the imaging surface exceeds a difined tolerance where an image captured with the photo-detector array 102 is deemed acceptable).
- a defined focal length e.g., a defect that would produce portion 108 of image 100 at some place behind an imaging surface of photo-detector 102 or a defect that would produce portion 104 at some place in front of the imaging surface of photo-detector array 102 where the distance in front or back of the imaging surface exceeds a difined tolerance where an image captured with the photo-detector array 102 is deemed acceptable
- the lens may have at least one spherical aberration that exceeds a first tolerance from a defined spherical aberration, or the lens may have at least one cylindrical aberration that exceeds a first tolerance from a defined cylindrical aberration. Alternatively, the lens may have some combination of such defects.
- method step 302 includes the sub-step of capturing the primary image at a primary focal surface location of the lens (e.g., a defined focal surface of the lens where an image would form if the lens had no aberrations).
- method step 302 includes the sub-step of capturing the primary image with a photo-detector array at the primary focal surface location of the lens (e.g., positioning the lens such that a defined focal surface of the lens coincides with an imaging surface of a photo-detector array).
- step 302 ( FIG. 3 ) would be controller 208 directing lens system 200 to position lens 204 at a primary position, and thereafter instructing image capture unit 100 to capture an image from photo-detector 102 .
- method step 304 illustrates determining at least one out-of-focus region of the primary image (or determining at least one focused region of the primary image).
- method step 304 includes the sub-step of calculating a Fourier transform of at least a part of the primary image (e.g., sharp, or in-focus images produce abrupt transitions that often have significant high frequency components).
- one specific example of method step 304 would be focus detection unit 210 performing a Fourier transform and subsequent analysis on at least a part of an image captured by image capture unit 206 when lens 204 was at the primary position.
- focus detection unit 210 could deem portions of the image having significant high frequency components as “in focus” images.
- method step 306 shows capturing another image with the lens at another position.
- method step 306 includes the sub-step of capturing the other image at the primary focal surface location of the lens at the primary position (e.g., lens 204 is moved to another position, while photo-detector 102 remains stationary, such as shown and described in relation to FIGS. 4 and 5 ).
- the step of capturing the other image at a primary focal surface location of the lens at the primary position further includes the sub-step of moving at least a part of the lens to the other position; and capturing the other image with a photo-detector array at the primary focal surface location of the lens at the primary position.
- the step of moving at least a part of the lens to the other position further includes the sub-step of moving the at least a part of the lens to the other position within at least one distance constrained by the first tolerance from the target optical property. In another implementation, the step of moving at least a part of the lens to the other position further includes the sub-step of moving an intermediary lens.
- the step of moving at least a part of the lens to the other position further includes the sub-step of distorting the lens such that the at least a part of the lens resides at the other position (e.g., a part of lens 204 is moved to another position, such as might happen if lens 204 were to be compressed laterally in a controlled manner, while photo-detector 102 remains stationary, such as shown and described in relation to FIGS. 4 and 5 ).
- FIG. 4 shows and describes moving at least a portion of the lens forward of the primary position (e.g., such as by controller 208 moving lens 204 forward, or causing lens 204 to be compressed such that a part of lens 204 moves forward relative to an imaging surface of photo-detector 102 ).
- FIG. 4 shows and describes moving at least a portion of the lens forward of the primary position (e.g., such as by controller 208 moving lens 204 forward, or causing lens 204 to be compressed such that a part of lens 204 moves forward relative to an imaging surface of photo-detector 102 ).
- FIG. 5 shows and describes moving at least a portion of the lens rearward of the primary position (e.g., such as by controller 208 moving lens 204 forward, or causing lens 204 to be compressed such that a part of lens 204 moves rearward relative to an imaging surface of photo-detector 102 ).
- method step 308 depicts determining a focus of at least one region of the other image relative to a focus of the at least one out-of-focus region of the primary image.
- method step 310 includes the sub-step of calculating a Fourier transform of at least a part of at least one region of the other image (e.g., sharp or in-focus images produce abrupt transitions that often have significant high frequency components).
- the step of calculating a Fourier transform of at least a part of at least one region of the other image includes the sub-step of mapping at least one region of the primary image with at least one region of the other image (e.g., mapping an out-of-focus region of the first image to a corresponding region of the second image).
- one specific example of method step 302 would be focus detection unit 210 performing a Fourier transform and subsequent analysis on at least a part of an image captured by image capture unit 206 when lens 204 was at the other position specified by controller 208 .
- method step 310 depicts constructing a composite image in response to the at least one region of the other image having a sharper focus relative to the focus of the at least one out-of-focus region of the primary image.
- the step of constructing a composite image in response to the at least one region of the other image having a sharper focus relative to the focus of the at least one out-of-focus region of the primary image includes the sub-step of replacing at least a part of the out-of-focus region of the primary image with at least a part of the at least one region of the other image.
- the step of constructing a composite image in response to the at least one region of the other image having a sharper focus relative to the focus of the at least one out-of-focus region of the primary image includes the sub-step of replacing at least a part of the out-of-focus region of the primary image with at least a part of the at least one region of the other image.
- the step of constructing a composite image in response to the at least one region of the other image having a sharper focus relative to the focus of the at least one out-of-focus region of the primary image includes the sub-step of utilizing at least one of tiling image processing techniques, morphing image processing techniques, blending image processing techniques, and stitching image processing techniques.
- the step of constructing a composite image in response to the at least one region of the other image having a sharper focus relative to the focus of the at least one out-of-focus region of the primary image includes the sub-steps of correlating a feature of the primary image with a feature of the other image; detecting at least one of size, color, and displacement distortion of at least one of the primary image and the other image; correcting the detected at least one of size, color, and displacement distortion of the at least one of the primary image and the other image; and assembling the composite using the corrected distortion.
- the step of constructing a composite image in response to the at least one region of the other image having a sharper focus relative to the focus of the at least one out-of-focus region of the primary image includes the sub-step of correcting for motion between the primary and the other image.
- one specific example of method step 302 would be image construction unit 212 creating a composite image by replacing those portions of an image of person 202 captured at a primary position with more in-focus positions of an image of person 202 captured by image capture unit 206 when lens 204 was at the other position.
- image construction unit 212 corrects for the motion between images using conventional techniques if such correction is desired. In another implementation of the example, motion correction is not used.
- method step 312 shows a determination of whether an aggregate change in focus, relative to the primary position of method step 302 , has exceeded a maximum expected deviation of a lens.
- a maximum expected deviation of a lens For example, even with a relatively poor quality lens, there will typically be an upper manufacturing limit beyond which lens defects are not expected to go (e.g., the lens has manufacturing criteria such as a focal length of 5 mm +/ ⁇ 0.05 mm).
- controller 208 comparing an aggregate movement in a defined direction against a pre-stored upper limit deviation value.
- controller 208 will determine whether the total forward movement of the lens is greater than 0.05 mm relative to the primary position.
- controller 208 will determine whether the total rearward movement of the lens is greater than 0.05 mm relative to the primary position.
- Method step 314 illustrates that the current composite image (e.g., of method step 310 ) is stored and/or displayed.
- One specific example of method step 314 would be store/display unit 214 either storing or displaying the composite image.
- Method step 316 shows the end of the process.
- step 312 shown is that in the event that the upper limit on lens tolerance has not been met or exceeded, the process proceeds to method step 306 and continues as described herein.
- FIG. 4 depicted is a side-plan view of the system of FIG. 2 wherein lens 204 has been moved in accordance with aspects of the process shown and described in relation to FIG. 3 .
- Lens 204 of lens system 200 is illustrated as having been moved to another position forward of the primary position which gave rise to the five different portions of image 100 shown and described in relation to FIGS. 1 and 2 .
- lens 204 of lens system 200 is illustrated as repositioned such that fifth portion 108 of image 100 is right sized and focused on an imaging surface of photo-detector array 102 (e.g., as shown and described in relation to method step 306 ).
- fifth portion 108 of image 100 can be combined with previously captured in focus and right sized portions 106 (e.g., FIGS. 1 and 2 ) to create a composite image such that the defects associated with fifth portion 108 as shown and described in relation to FIGS. 1 and 2 are alleviated (e.g., as shown and described in relation to method step 310 ).
- the remaining components and control aspects of the various parts of FIG. 4 function as described elsewhere herein.
- FIG. 5 illustrated is another side-plan view of the system of FIG. 2 wherein lens 204 has been moved in accordance with aspects of the process shown and described in relation to FIG. 3 .
- Lens 204 of lens system 200 is illustrated as having been moved to another position rearward of the primary position which gave rise to the five different portions of image 100 shown and described in relation to FIG. 1 .
- lens 204 of lens system 200 is illustrated as positioned such that first portion 104 of image 100 is right sized and focused on an imaging surface of photo-detector array 102 (e.g., as described in relation to method step 306 ).
- first portion 104 of image 100 can be combined with previously captured in focus and right sized portions 106 , 108 (e.g., FIGS. 1, 2 , and 4 ) to create a composite image such that the defects associated with first portion 104 as shown and described in relation to FIGS. 1 and 2 are alleviated (e.g., as shown and described in relation to method step 310 ).
- the remaining components and control aspects of the various parts of FIG. 5 function as described elsewhere herein.
- an implementer may opt for a hardware and/or firmware vehicle; alternatively, if flexibility is paramount, the implementer may opt for a solely software implementation; or, yet again alternatively, the implementer may opt for some combination of hardware, software, and/or firmware.
- any vehicle to be utilized is a choice dependent upon the context in which the vehicle will be deployed and the specific concerns (e g., speed, flexibility, or predictability) of the implementer, any of which may vary.
- Those skilled in the art will recognize that optical aspects of implementations will require optically-oriented hardware, software, and or firmware.
- a signal bearing media include, but are not limited to, the following: recordable type media such as floppy disks, hard disk drives, CD ROMs, digital tape, and computer memory; and transmission type media such as digital and analog communication links using TDM or IP based communication links (e.g., packet links).
- electro-mechanical system includes, but is not limited to, electrical circuitry operably coupled with a transducer (e.g., an actuator, a motor, a piezoelectric crystal, etc.), electrical circuitry having at least one discrete electrical circuit, electrical circuitry having at least one integrated circuit, electrical circuitry having at least one application specific integrated circuit, electrical circuitry forming a general purpose computing device configured by a computer program (e.g., a general purpose computer configured by a computer program which at least partially carries out processes and/or devices described herein, or a microprocessor configured by a computer program which at least partially carries out processes and/or devices described herein), electrical circuitry forming a memory device (e.g., forms of random access memory), electrical circuitry forming a communications device (e.g., a modem, communications switch, or optical-electrical equipment), and any non-electrical analog thereto, such as optical re other analogs.
- a transducer e.g., an actuator, a motor, a piez
- electro-mechanical systems include but are not limited to a variety of consumer electronics systems, as well as other systems such as motorized transport systems, factory automation systems, security systems, and communication/computing systems.
- electro-mechanical as used herein is not necessarily limited to a system that has both electrical and mechanical actuation except as context may dictate otherwise.
- a typical image processing system generally includes one or more of a system unit housing, a video display device, a memory such as volatile and non-volatile memory, processors such as microprocessors and digital signal processors, computational entities such as operating systems, drivers, and applications programs, one or more interaction devices, such as a touch pad or screen, control systems including feedback loops and control motors (e.g., feedback for sensing lens position and/or velocity; control motors for moving/distorting lenses to give desired focuses.
- a typical image processing system may be implemented utilizing any suitable commercially available components, such as those typically found in digital still systems and/or digital motion systems.
- any two components herein combined to achieve a particular functionality can be seen as “associated with” each other such that the desired functionality is achieved, irrespective of architectures or intermedial components.
- any two components so associated can also be viewed as being “operably connected” or “operably coupled” to each other to achieve the desired functionality.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Studio Devices (AREA)
- Image Processing (AREA)
Abstract
A system constructs a composite image using focus assessment information of image regions.
Description
- The present application relates, in general, to imaging.
- In one embodiment, a method includes but is not limited to: capturing a primary image with a lens at a primary position, the lens having at least one deviation that exceeds a first tolerance from a target optical property; capturing another image with the lens at another position; determining at least one out-of-focus region of the primary image; determining a focus of at least one region of the other image relative to a focus of the at least one out-of-focus region of the primary image; and constructing a composite image in response to the at least one region of the other image having a sharper focus relative to the focus of the at least one out-of-focus region of the primary image. In addition to the foregoing, various other method embodiments are set forth and described in the text (e.g., claims and/or detailed description) and/or drawings of the present application.
- In one or more various embodiments, related systems include but are not limited to electro-mechanical systems (e.g., motors, actuators, circuitry, and/or programming) for effecting the herein-referenced method embodiments); the electrical circuitry can be virtually any combination of hardware, software, and/or firmware configured to effect the foregoing- referenced method embodiments depending upon the design choices of the system designer.
- In one embodiment, a system includes but is not limited to: a photo-detector array; a lens having at least one deviation that exceeds a first tolerance from a target optical property; a controller configured to position said lens at a primary and another position relative to said photo-detector array and to cause an image capture signal at the primary and the other position; and an image construction unit configured to construct at least one out-of-focus region of a first image captured at the primary position with a more in-focus region of another image captured at the other position.
- In addition to the foregoing, various other method and or system embodiments are set forth and described in the text (e.g., claims and/or detailed description) and/or drawings of the present application.
- The foregoing is a summary and thus contains, by necessity; simplifications, generalizations and omissions of detail; consequently, those skilled in the art will appreciate that the summary is illustrative only and is NOT intended to be in any way limiting. Other aspects, inventive features, and advantages of the devices and/or processes described herein, as defined solely by the claims, will become apparent in the non-limiting detailed description set forth herein.
-
FIG. 1 shows a front-plan view ofimage 100 of a person (e.g.,person 202 ofFIG. 2 ) projected onto photo-detector array 102. -
FIG. 2 depicts a side-plan view oflens system 200 that can give rise toimage 100 ofFIG. 1 . -
FIG. 3 depicts a high level logic flowchart of a process. -
FIG. 4 depicts a side-plan view of the system ofFIG. 2 whereinlens 204 has been moved in accordance with aspects of the process shown and described in relation toFIG. 3 . -
FIG. 5 illustrates another side-plan view of the system ofFIG. 2 whereinlens 204 has been moved in accordance with aspects of the process shown and described in relation toFIG. 3 . - The use of the same symbols in different drawings typically indicates similar or identical items.
- With reference to the figures, and with reference now to
FIG. 1 , shown is a front-plan view ofimage 100 of a person (e.g.,person 202 ofFIG. 2 ) projected onto photo-detector array 102.Image 100 is shown as distorted due to defects in a lens through whichimage 100 has been projected (e.g.,lens 204 oflens system 200 ofFIG. 2 ).First portion 104 ofimage 100 is illustrated as large and blurry, which can occur when a lens defect causesportion 104 ofimage 100 to come to a focus in front of a surface of photo-detector array 102. Second, third, andfourth portions 106 are illustrated as right sized, which can occur when the lens causesportions 106 ofimage 100 to correctly focus on an imaging surface of photo-detector array 102.Fifth portion 108 is shown as small and faint, which can occur when a lens defect causesportion 108 ofimage 100 to come to a focus (virtual) behind an imaging surface of photo-detector array 102. In addition, although not expressly shown, those having skill in the art will appreciate that various lens defects could also cause the image to be distorted in x-y; those having skill in the art will also appreciate that different colored wavelengths of light can in and of themselves focus at different positions due to differences in refraction of the different colored wavelengths of light. - Referring now to
FIG. 2 , depicted is a side-plan view oflens system 200 that can give rise toimage 100 ofFIG. 1 .Lens 204 oflens system 200 is illustrated as located at a primary position and having defects that give rise to the five different portions ofimage 100 shown and described in relation toFIG. 1 .First portion 104 ofimage 100 is illustrated as focused in front of an imaging surface of photo-detector array 102. Second, third, andfourth portions 106 are illustrated as right sized and focused on an imaging surface of photo-detector array 102. (It is recognized that in side plan view the head and feet ofperson 202 would appear as lines; however, for sake of clarity they are shown in profile inFIG. 2 to help orient the reader relative toFIG. 1 .)Fifth portion 108 is shown as small and faint, and virtually focused behind an imaging surface of photo-detector array 102. - Continuing to refer to
FIG. 2 , further shown are components that can serve as the environment for the process shown and described in relation toFIG. 3 . Specifically,controller 208 is depicted as controlling the position oflens 204 of lens system 200 (e.g., via use of a feedback control subsystem).Image capture unit 206 is illustrated as receiving image data from photo-detector 102 and receiving control signals fromcontroller 208.Image capture unit 206 is shown as transmitting captured image information to focusdetection unit 210.Focus detection unit 210 is depicted as transmitting focus data toimage construction unit 212.Image construction unit 212 is illustrated as transmitting a composite image to image store/display unit 214. - With reference now to
FIG. 3 , depicted is a high level logic flowchart of a process.Method step 300 shows the start of the process.Method step 302 depicts capturing a primary image with a lens at a primary position, the lens having at least one deviation that exceeds a first tolerance from a target optical property. One example of the lens having at least one deviation that exceeds a first tolerance from a target optical property would be where the lens has at least one focal length that exceeds a first tolerance from a defined focal length (e.g., a defect that would produceportion 108 ofimage 100 at some place behind an imaging surface of photo-detector 102 or a defect that would produceportion 104 at some place in front of the imaging surface of photo-detector array 102 where the distance in front or back of the imaging surface exceeds a difined tolerance where an image captured with the photo-detector array 102 is deemed acceptable). For instance, the lens may have at least one spherical aberration that exceeds a first tolerance from a defined spherical aberration, or the lens may have at least one cylindrical aberration that exceeds a first tolerance from a defined cylindrical aberration. Alternatively, the lens may have some combination of such defects. In one implementation,method step 302 includes the sub-step of capturing the primary image at a primary focal surface location of the lens (e.g., a defined focal surface of the lens where an image would form if the lens had no aberrations). In another implementation,method step 302 includes the sub-step of capturing the primary image with a photo-detector array at the primary focal surface location of the lens (e.g., positioning the lens such that a defined focal surface of the lens coincides with an imaging surface of a photo-detector array). - Referring again to
FIG. 2 , one specific example of method step 302 (FIG. 3 ) would becontroller 208directing lens system 200 to positionlens 204 at a primary position, and thereafter instructingimage capture unit 100 to capture an image from photo-detector 102. - With reference again to
FIG. 3 ,method step 304 illustrates determining at least one out-of-focus region of the primary image (or determining at least one focused region of the primary image). In one implementation,method step 304 includes the sub-step of calculating a Fourier transform of at least a part of the primary image (e.g., sharp, or in-focus images produce abrupt transitions that often have significant high frequency components). - Referring again to
FIG. 2 , one specific example of method step 304 (FIG. 3 ) would befocus detection unit 210 performing a Fourier transform and subsequent analysis on at least a part of an image captured byimage capture unit 206 whenlens 204 was at the primary position. In this example,focus detection unit 210 could deem portions of the image having significant high frequency components as “in focus” images. - With reference again to
FIG. 3 ,method step 306 shows capturing another image with the lens at another position. In one implementation,method step 306 includes the sub-step of capturing the other image at the primary focal surface location of the lens at the primary position (e.g.,lens 204 is moved to another position, while photo-detector 102 remains stationary, such as shown and described in relation toFIGS. 4 and 5 ). In another implementation, the step of capturing the other image at a primary focal surface location of the lens at the primary position further includes the sub-step of moving at least a part of the lens to the other position; and capturing the other image with a photo-detector array at the primary focal surface location of the lens at the primary position. In another implementation, the step of moving at least a part of the lens to the other position further includes the sub-step of moving the at least a part of the lens to the other position within at least one distance constrained by the first tolerance from the target optical property. In another implementation, the step of moving at least a part of the lens to the other position further includes the sub-step of moving an intermediary lens. In another implementation, the step of moving at least a part of the lens to the other position further includes the sub-step of distorting the lens such that the at least a part of the lens resides at the other position (e.g., a part oflens 204 is moved to another position, such as might happen iflens 204 were to be compressed laterally in a controlled manner, while photo-detector 102 remains stationary, such as shown and described in relation toFIGS. 4 and 5 ). - Referring now to
FIGS. 2, 4 and/or 5, one specific example of method step 306 (FIG. 3 ) would becontroller 208directing lens system 200 to positionlens 204 at a position other than the primary position and thereafter instructingimage capture unit 100 to capture an image from photo-detector 102.FIG. 4 shows and describes moving at least a portion of the lens forward of the primary position (e.g., such as bycontroller 208 movinglens 204 forward, or causinglens 204 to be compressed such that a part oflens 204 moves forward relative to an imaging surface of photo-detector 102).FIG. 5 shows and describes moving at least a portion of the lens rearward of the primary position (e.g., such as bycontroller 208 movinglens 204 forward, or causinglens 204 to be compressed such that a part oflens 204 moves rearward relative to an imaging surface of photo-detector 102). - With reference again to
FIG. 3 ,method step 308 depicts determining a focus of at least one region of the other image relative to a focus of the at least one out-of-focus region of the primary image. In one implementation,method step 310 includes the sub-step of calculating a Fourier transform of at least a part of at least one region of the other image (e.g., sharp or in-focus images produce abrupt transitions that often have significant high frequency components). In one implementation, the step of calculating a Fourier transform of at least a part of at least one region of the other image (e.g., sharp or in-focus images produce abrupt transitions that often have significant high frequency components) includes the sub-step of mapping at least one region of the primary image with at least one region of the other image (e.g., mapping an out-of-focus region of the first image to a corresponding region of the second image). - Referring again to
FIGS. 2, 4 and/or 5, one specific example of method step 302 (FIG. 3 ) would befocus detection unit 210 performing a Fourier transform and subsequent analysis on at least a part of an image captured byimage capture unit 206 whenlens 204 was at the other position specified bycontroller 208. - With reference again to
FIG. 3 ,method step 310 depicts constructing a composite image in response to the at least one region of the other image having a sharper focus relative to the focus of the at least one out-of-focus region of the primary image. In one implementation, the step of constructing a composite image in response to the at least one region of the other image having a sharper focus relative to the focus of the at least one out-of-focus region of the primary image includes the sub-step of replacing at least a part of the out-of-focus region of the primary image with at least a part of the at least one region of the other image. In another implementation, the step of constructing a composite image in response to the at least one region of the other image having a sharper focus relative to the focus of the at least one out-of-focus region of the primary image includes the sub-step of replacing at least a part of the out-of-focus region of the primary image with at least a part of the at least one region of the other image. In yet another implementation, the step of constructing a composite image in response to the at least one region of the other image having a sharper focus relative to the focus of the at least one out-of-focus region of the primary image includes the sub-step of utilizing at least one of tiling image processing techniques, morphing image processing techniques, blending image processing techniques, and stitching image processing techniques. - In yet another implementation, the step of constructing a composite image in response to the at least one region of the other image having a sharper focus relative to the focus of the at least one out-of-focus region of the primary image includes the sub-steps of correlating a feature of the primary image with a feature of the other image; detecting at least one of size, color, and displacement distortion of at least one of the primary image and the other image; correcting the detected at least one of size, color, and displacement distortion of the at least one of the primary image and the other image; and assembling the composite using the corrected distortion. In yet another implementation, the step of constructing a composite image in response to the at least one region of the other image having a sharper focus relative to the focus of the at least one out-of-focus region of the primary image includes the sub-step of correcting for motion between the primary and the other image.
- Referring again to
FIGS. 2, 4 and/or 5, one specific example of method step 302 (FIG. 3 ) would beimage construction unit 212 creating a composite image by replacing those portions of an image ofperson 202 captured at a primary position with more in-focus positions of an image ofperson 202 captured byimage capture unit 206 whenlens 204 was at the other position. In one implementation of the example,image construction unit 212 corrects for the motion between images using conventional techniques if such correction is desired. In another implementation of the example, motion correction is not used. - With reference again to
FIG. 3 ,method step 312 shows a determination of whether an aggregate change in focus, relative to the primary position ofmethod step 302, has exceeded a maximum expected deviation of a lens. For example, even with a relatively poor quality lens, there will typically be an upper manufacturing limit beyond which lens defects are not expected to go (e.g., the lens has manufacturing criteria such as a focal length of 5 mm +/−0.05 mm). - Referring again to
FIGS. 2, 4 and/or 5, one specific example of method step 312 (FIG. 3 ) would becontroller 208 comparing an aggregate movement in a defined direction against a pre-stored upper limit deviation value. In an implementation of the example illustrated inFIG. 4 , iflens 204 has manufacturing criteria such as a focal length of 5 mm +/−0.05 mm,controller 208 will determine whether the total forward movement of the lens is greater than 0.05 mm relative to the primary position. In an implementation of the example illustrated inFIG. 5 , iflens 204 has manufacturing criteria such as a focal length of 5 mm +/−0.05 mm,controller 208 will determine whether the total rearward movement of the lens is greater than 0.05 mm relative to the primary position. - With reference again to
FIG. 3 , if the inquiry ofmethod step 312 yields a determination that the aggregate changes in focuses has met or exceeded the maximum expected deviation of the lens, the process proceeds tomethod step 314.Method step 314 illustrates that the current composite image (e.g., of method step 310) is stored and/or displayed. One specific example ofmethod step 314 would be store/display unit 214 either storing or displaying the composite image. -
Method step 316 shows the end of the process. - Returning to
method step 312, shown is that in the event that the upper limit on lens tolerance has not been met or exceeded, the process proceeds tomethod step 306 and continues as described herein. - Referring now to
FIG. 4 , depicted is a side-plan view of the system ofFIG. 2 whereinlens 204 has been moved in accordance with aspects of the process shown and described in relation toFIG. 3 .Lens 204 oflens system 200 is illustrated as having been moved to another position forward of the primary position which gave rise to the five different portions ofimage 100 shown and described in relation toFIGS. 1 and 2 . Specifically,lens 204 oflens system 200 is illustrated as repositioned such thatfifth portion 108 ofimage 100 is right sized and focused on an imaging surface of photo-detector array 102 (e.g., as shown and described in relation to method step 306). In one implementation,fifth portion 108 ofimage 100 can be combined with previously captured in focus and right sized portions 106 (e.g.,FIGS. 1 and 2 ) to create a composite image such that the defects associated withfifth portion 108 as shown and described in relation toFIGS. 1 and 2 are alleviated (e.g., as shown and described in relation to method step 310). The remaining components and control aspects of the various parts ofFIG. 4 function as described elsewhere herein. - With reference now to
FIG. 5 , illustrated is another side-plan view of the system ofFIG. 2 whereinlens 204 has been moved in accordance with aspects of the process shown and described in relation toFIG. 3 .Lens 204 oflens system 200 is illustrated as having been moved to another position rearward of the primary position which gave rise to the five different portions ofimage 100 shown and described in relation toFIG. 1 . Specifically,lens 204 oflens system 200 is illustrated as positioned such thatfirst portion 104 ofimage 100 is right sized and focused on an imaging surface of photo-detector array 102 (e.g., as described in relation to method step 306). In one implementation,first portion 104 ofimage 100 can be combined with previously captured in focus and rightsized portions 106, 108 (e.g.,FIGS. 1, 2 , and 4) to create a composite image such that the defects associated withfirst portion 104 as shown and described in relation toFIGS. 1 and 2 are alleviated (e.g., as shown and described in relation to method step 310). The remaining components and control aspects of the various parts ofFIG. 5 function as described elsewhere herein. - Those having skill in the art will recognize that the state of the art has progressed to the point where there is little distinction left between hardware and software implementations of aspects of systems; the use of hardware or software is generally (but not always, in that in certain contexts the choice between hardware and software can become significant) a design choice representing cost vs. efficiency tradeoffs. Those having skill in the art will appreciate that there are various vehicles by which processes and/or systems described herein can be effected (e.g., hardware, software, and/or firmware), and that the preferred vehicle will vary with the context in which the processes are deployed. For example, if an implementer determines that speed and accuracy are paramount, the implementer may opt for a hardware and/or firmware vehicle; alternatively, if flexibility is paramount, the implementer may opt for a solely software implementation; or, yet again alternatively, the implementer may opt for some combination of hardware, software, and/or firmware. Hence, there are several possible vehicles by which the processes described herein may be effected, none of which is inherently superior to the other in that any vehicle to be utilized is a choice dependent upon the context in which the vehicle will be deployed and the specific concerns (e g., speed, flexibility, or predictability) of the implementer, any of which may vary. Those skilled in the art will recognize that optical aspects of implementations will require optically-oriented hardware, software, and or firmware.
- The foregoing detailed description has set forth various embodiments of the devices and/or processes via the use of block diagrams, flowcharts, and examples. Insofar as such block diagrams, flowcharts, and examples contain one or more functions and/or operations, it will be understood as notorious by those within the art that each function and/or operation within such block diagrams, flowcharts, or examples can be implemented, individually and/or collectively, by a wide range of hardware, software, firmware, or virtually any combination thereof. In one embodiment, the present invention may be implemented via Application Specific Integrated Circuits (ASICs), Field Programmable Gate Arrays (FPGAs), or other integrated formats. However, those skilled in the art will recognize that the embodiments disclosed herein, in whole or in part, can be equivalently implemented in standard integrated circuits, as one or more computer programs running on one or more computers (e.g., as one or more programs running on one or more computer systems), as one or more programs running on one or more processors (e.g., as one or more programs running on one or more microprocessors), as firmware, or as virtually any combination thereof, and that designing the circuitry and/or writing the code for the software and or firmware would be well within the skill of one of skill in the art in light of this disclosure. In addition, those skilled in the art will appreciate that the mechanisms of the present invention are capable of being distributed as a program product in a variety of forms, and that an illustrative embodiment of the present invention applies equally regardless of the particular type of signal bearing media used to actually carry out the distribution. Examples of a signal bearing media include, but are not limited to, the following: recordable type media such as floppy disks, hard disk drives, CD ROMs, digital tape, and computer memory; and transmission type media such as digital and analog communication links using TDM or IP based communication links (e.g., packet links).
- In a general sense, those skilled in the art will recognize that the various embodiments described herein which can be implemented, individually and/or collectively, by various types of electro-mechanical systems having a wide range of electrical components such as hardware, software, firmware, or virtually any combination thereof; and a wide range of components that may impart mechanical force or motion such as rigid bodies, spring or torsional bodies, hydraulics, and electro-magnetically actuated devices, or virtually any combination thereof. Consequently, as used herein “electro-mechanical system” includes, but is not limited to, electrical circuitry operably coupled with a transducer (e.g., an actuator, a motor, a piezoelectric crystal, etc.), electrical circuitry having at least one discrete electrical circuit, electrical circuitry having at least one integrated circuit, electrical circuitry having at least one application specific integrated circuit, electrical circuitry forming a general purpose computing device configured by a computer program (e.g., a general purpose computer configured by a computer program which at least partially carries out processes and/or devices described herein, or a microprocessor configured by a computer program which at least partially carries out processes and/or devices described herein), electrical circuitry forming a memory device (e.g., forms of random access memory), electrical circuitry forming a communications device (e.g., a modem, communications switch, or optical-electrical equipment), and any non-electrical analog thereto, such as optical re other analogs. Those skilled in the art will also appreciate that examples of electro-mechanical systems include but are not limited to a variety of consumer electronics systems, as well as other systems such as motorized transport systems, factory automation systems, security systems, and communication/computing systems. Those skilled in the art will recognize that electro-mechanical as used herein is not necessarily limited to a system that has both electrical and mechanical actuation except as context may dictate otherwise.
- Those skilled in the art will recognize that it is common within the art to describe devices and/or processes in the fashion set forth herein, and thereafter use standard engineering practices to integrate such described devices and/or processes into image processing systems. That is, at least a portion of the devices and/or processes described herein can be integrated into an image processing system via a reasonable amount of experimentation. Those having skill in the art will recognize that a typical image processing system generally includes one or more of a system unit housing, a video display device, a memory such as volatile and non-volatile memory, processors such as microprocessors and digital signal processors, computational entities such as operating systems, drivers, and applications programs, one or more interaction devices, such as a touch pad or screen, control systems including feedback loops and control motors (e.g., feedback for sensing lens position and/or velocity; control motors for moving/distorting lenses to give desired focuses. A typical image processing system may be implemented utilizing any suitable commercially available components, such as those typically found in digital still systems and/or digital motion systems.
- The foregoing described embodiments depict different components contained within, or connected with, different other components. It is to be understood that such depicted architectures are merely exemplary, and that in fact many other architectures can be implemented which achieve the same functionality. In a conceptual sense, any arrangement of components to achieve the same functionality is effectively “associated” such that the desired functionality is achieved. Hence, any two components herein combined to achieve a particular functionality can be seen as “associated with” each other such that the desired functionality is achieved, irrespective of architectures or intermedial components. Likewise, any two components so associated can also be viewed as being “operably connected” or “operably coupled” to each other to achieve the desired functionality.
- While particular embodiments of the present invention have been shown and described, it will be understood by those skilled in the art that, based upon the teachings herein, changes and modifications may be made without departing from this invention and its broader aspects and, therefore, the appended claims are to encompass within their scope all such changes and modifications as are within the true spirit and scope of this invention. Furthermore, it is to be understood that the invention is solely defined by the appended claims. It will be understood by those within the art that, in general, terms used herein, and especially in the appended claims (e.g., bodies of the appended claims) are generally intended as “open” terms (e.g., the term “including” should be interpreted as “including but not limited to,” the term “having” should be interpreted as “having at least,” the term “includes” should be interpreted as “includes but is not limited to,”“comprise” and variations thereof, such as, “comprises” and “comprising” are to be construed in an open, inclusive sense, that is as “including, but not limited to,” etc.). It will be further understood by those within the art that if a specific number of an introduced claim recitation is intended, such an intent will be explicitly recited in the claim, and in the absence of such recitation no such intent is present. For example, as an aid to understanding, the following appended claims may contain usage of the introductory phrases “at least one” and “one or more” to introduce claim recitations. However, the use of such phrases should not be construed to imply that the introduction of a claim recitation by the indefinite articles “a” or “an” limits any particular claim containing such introduced claim recitation to inventions containing only one such recitation, even when the same claim includes the introductory phrases “one or more” or “at least one” and indefinite articles such as “a” or “an” (e.g., “a” and/or “an” should typically be interpreted to mean “at least one” or “one or more”); the same holds true for the use of definite articles used to introduce claim recitations. In addition, even if a specific number of an introduced claim recitation is explicitly recited, those skilled in the art will recognize that such recitation should typically be interpreted to mean at least the recited number (e.g., the bare recitation of “two recitations,” without other modifiers, typically means at least two recitations, or two or more recitations).
Claims (56)
1. A first method comprising:
performing a reception of or a transmission of one or more instructions in relation to a second method that includes at least:
capturing a primary image with a lens at a primary position, the lens having at least one deviation that exceeds a first tolerance from a target optical property;
capturing another image with the lens at another position;
determining at least one out-of-focus region of the primary image;
determining a focus of at least one region of the other image relative to a focus of the at least one out-of-focus region of the primary image; and
constructing a composite image in response to the at least one region of the other image having a sharper focus relative to the focus of the at least one out-of-focus region of the primary image.
2. (canceled)
3. (canceled)
4. (canceled)
5. (canceled)
6. (canceled)
7. (canceled)
8. (canceled)
9. (canceled)
10. (canceled)
11. (canceled)
12. (canceled)
13. (canceled)
14. (canceled)
15. (canceled)
16. (canceled)
17. (canceled)
18. (canceled)
19. A first system comprising:
means for performing a reception of or a transmission of one or more instructions in relation to a second system that includes at least:
means for capturing a primary image with a lens at a primary position, the lens having at least one deviation that exceeds a first tolerance from a target optical property;
means for capturing another image with the lens at another position;
means for determining at least one out-of-focus region of the primary image;
means for determining a focus of at least one region of the other image relative to a focus of the at least one out-of-focus region of the primary image; and
means for constructing a composite image in response to the at least one region of the other image having a sharper focus relative to the focus of the at least one out-of-focus region of the primary image.
20. The second system of claim 19 , wherein said lens having at least one deviation that exceeds a first tolerance from a target optical property further comprises:
the lens having at least one focal length that exceeds a first tolerance from a defined focal length.
21. The second system of claim 19 , wherein said lens having at least one deviation that exceeds a first tolerance from a target optical property further comprises:
the lens having at least one spherical aberration that exceeds a first tolerance from a defined spherical aberration.
22. The second system of claim 19 , wherein said lens having at least one deviation that exceeds a first tolerance from a target optical property further comprises:
the lens having at least one cylindrical aberration that exceeds a first tolerance from a defined cylindrical aberration.
23. The second system of claim 19 , wherein said means for capturing a primary image with a lens at a primary position further comprises:
means for capturing the primary image at a primary focal surface location of the lens.
24. The second system of claim 23 , wherein said means for capturing the primary image at a primary focal surface location of the lens further comprises:
means for capturing the primary image with a photo-detector array at the primary focal surface location of the lens.
25. The second system of claim 19 , wherein said means for capturing another image with the lens at another position further comprises:
means for capturing the other image at a primary focal surface location of the lens at the primary position.
26. The second system of claim 25 , wherein said means for capturing the other image at a primary focal surface location of the lens at the primary position further comprises:
means for moving at least a part of the lens to the other position; and
means for capturing the other image with a photo-detector array at the primary focal surface location of the lens at the primary position.
27. The second system of claim 26 , wherein said means for moving at least a part of the lens to the other position further comprises:
means for moving the at least a part of the lens to the other position within at least one distance constrained by the first tolerance from the target optical property.
28. The second system of claim 26 , wherein said means for moving at least a part of the lens to the other position further comprises:
means for moving an intermediary lens.
29. The second system of claim 26 , wherein said means for moving at least a part of the lens to the other position further comprises:
means for distorting the lens such that the at least a part of the lens resides at the other position.
30. The second system of claim 19 , wherein said means for determining at least one out-of-focus region of the primary image further comprises:
means for calculating a Fourier transform of at least a part of the primary image.
31. The second system of claim 19 , wherein said means for determining a focus of at least one region of the other image relative to a focus of the at least one out-of-focus region of the primary image further comprises:
means for calculating a Fourier transform of at least a part of the at least one region of the other image.
32. The second system of claim 31 , wherein said means for calculating a Fourier transform of at least a part of the at least one region of the other image further comprises:
means for mapping at least one region of the primary image with at least one region of the other image.
33. The second system of claim 19 , wherein said means for constructing a composite image in response to the at least one region of the other image having a sharper focus relative to the focus of the at least one out-of-focus region of the primary image further comprises:
means for replacing at least a part of the out-of-focus region of the primary image with at least a part of the at least one region of the other image.
34. (canceled)
35. The second system of claim 19 , wherein said means for constructing a composite image in response to the at least one region of the other image having a sharper focus relative to the focus of the at least one out-of-focus region of the primary image further comprises:
means for correlating a feature of the primary image with a feature of the other image;
means for detecting at least one of size, color, and/or displacement distortion of at least one of the primary image and/or the other image;
means for correcting the detected at least one of size, color, and/or displacement distortion of the at least one of the primary image and/or the other image; and
means for assembling the composite image using the corrected distortion.
36. The second system of claim 19 , wherein said means for constructing a composite image in response to the at least one region of the other image having a sharper focus relative to the focus of the at least one out-of-focus region of the primary image further comprises:
means for correcting for motion between the primary and the other image.
37. A system comprising:
a photo-detector array;
a lens having at least one deviation that exceeds a first tolerance from a target optical property;
a controller configured to position said lens at a primary and another position relative to said photo-detector array and to cause an image capture signal at the primary and the other position; and
an image construction unit configured to construct at least one out-of-focus region of a first image captured at the primary position with a more in-focus region of another image captured at the other position.
38. The system of claim 37 , wherein said controller configured to position said lens at a primary and another position relative to said photo-detector array and to cause an image capture signal at the primary and the other position further comprises:
a transducer system having a control signal input operably coupled with said controller and a motion output operably coupled with said lens.
39. The system of claim 38 , wherein said transducer system further comprises an electric motor operably coupled to move said lens
40. The system of claim 38 , wherein said transducer system further comprises an electric motor operably coupled to distort said lens
41. The system of claim 37 , wherein said image construction unit configured to construct at least one out-of-focus region of a first image captured at the primary position with a more in-focus region of another image captured at the other position further comprises:
circuitry for constructing at least one out-of-focus region of a first image captured at the primary position with a more in-focus region of another image captured at the other position said circuitry including at least one of electrical circuitry having at least one discrete electrical circuit, electrical circuitry having at least one integrated circuit, electrical circuitry having at least one application specific integrated circuit, electrical circuitry having a general purpose computing device configured by a computer program, electrical circuitry having a memory device, and/or electrical circuitry having a communications device.
42. A system comprising:
a lens having at least one deviation that exceeds a first tolerance from a target optical property;
an electro-mechanical system configurable to capture a primary image with a lens at a primary position said electro-mechanical system including at least one of electrical circuitry operably coupled with a transducer, electrical circuitry having at least one discrete electrical circuit, electrical circuitry having at least one integrated circuit, electrical circuitry having at least one application specific integrated circuit, electrical circuitry having a general purpose computing device configured by a computer program, electrical circuitry having a memory device, and/or electrical circuitry having a communications device;
an electro-mechanical system configurable to capture another image with the lens at another position said electro-mechanical system including at least one of electrical circuitry operably coupled with a transducer, electrical circuitry having at least one discrete electrical circuit, electrical circuitry having at least one integrated circuit, electrical circuitry having at least one application specific integrated circuit, electrical circuitry having a general purpose computing device configured by a computer program, electrical circuitry having a memory device, and/or electrical circuitry having a communications device;
an electro-mechanical system configurable to determine at least one out-of-focus region of the primary image said electro-mechanical system including at least one of electrical circuitry operably coupled with a transducer, electrical circuitry having at least one discrete electrical circuit, electrical circuitry having at least one integrated circuit, electrical circuitry having at least one application specific integrated circuit, electrical circuitry having a general purpose computing device configured by a computer program, electrical circuitry having a memory device, and/or electrical circuitry having a communications device;
an electro-mechanical system configurable to determine a focus of at least one region of the other image relative to a focus of the at least one out-of-focus region of the primary image said electro-mechanical system including at least one of electrical circuitry operably coupled with a transducer, electrical circuitry having at least one discrete electrical circuit, electrical circuitry having at least one integrated circuit, electrical circuitry having at least one application specific integrated circuit, electrical circuitry having a general purpose computing device configured by a computer program, electrical circuitry having a memory device, and/or electrical circuitry having a communications device; and
an electro-mechanical system configurable to construct a composite image in response to the at least one region of the other image having a sharper focus relative to the focus of the at least one out-of-focus region of the primary image said electro-mechanical system including at least one of electrical circuitry operably coupled with a transducer, electrical circuitry having at least one discrete electrical circuit, electrical circuitry having at least one integrated circuit, electrical circuitry having at least one application specific integrated circuit, electrical circuitry having a general purpose computing device configured by a computer program, electrical circuitry having a memory device, and/or electrical circuitry having a communications device.
43. (canceled)
44. (canceled)
45. (canceled)
46. (canceled)
47. (canceled)
48. The first method of claim 1 wherein the performing a reception of or a transmission of one or more instructions in relation to a second method comprises:
transmitting at least one instruction representative of a patch generated in response to a representation of the second method.
49. The first method of claim 1 wherein the performing a reception of or a transmission of one or more instructions in relation to a second method comprises:
transmitting at least one instruction representative of an upgrade generated in response to a representation of the second method.
50. The first system of claim 19 further comprising:
means for receiving a user authorization for the performing the reception of or the transmission of the one or more instructions in relation to the second system.
51. The first system of claim 19 wherein the means for performing a reception of or a transmission of one or more instructions in relation to a second system comprises:
means for receiving the one or more instructions; and
means for replacing a portion of a representation of the second system in response to the one or more instructions.
52. The first system of claim 19 wherein the means for performing a reception of or a transmission of one or more instructions in relation to a second system comprises:
means for receiving the one or more instructions; and
means for patching a representation of the second system in response to the one or more instructions.
53. The first system of claim 19 wherein the means for performing a reception of or a transmission of one or more instructions in relation to a second system comprises:
means for receiving the one or more instructions; and
means for forming a representation of the second system in response to the one or more instructions.
54. The first system of claim 19 wherein the means for performing a reception of or a transmission of one or more instructions in relation to a second system comprises:
means for transmitting at least one indicator representative of the second system.
55. The first system of claim 19 wherein the means for performing a reception of or a transmission of one or more instructions in relation to a second system comprises:
means for transmitting at least one instruction representative of a patch generated in response to a representation of the second system.
56. The first system of claim 19 wherein the means for performing a reception of or a transmission of one or more instructions in relation to a second system comprises:
means for transmitting at least one instruction representative of an upgrade generated in response to a representation of the second system.
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/804,314 US20070280550A1 (en) | 2003-12-16 | 2007-05-15 | Lens defect correction |
US12/925,848 US8643955B2 (en) | 2003-12-16 | 2010-10-28 | Image correction using individual manipulation of microlenses in a microlens array |
US14/108,003 US9348123B2 (en) | 2003-12-16 | 2013-12-16 | Image correction using individual manipulation of microlenses in a microlens array |
US15/159,517 US20160343119A1 (en) | 2003-12-16 | 2016-05-19 | Selective Region-Based Focus With Focal Adjustment Bracketing Via Lens / Image Sensor Position Manipulation |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/738,626 US7231097B2 (en) | 2003-12-16 | 2003-12-16 | Lens defect correction |
US11/804,314 US20070280550A1 (en) | 2003-12-16 | 2007-05-15 | Lens defect correction |
Related Parent Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/738,626 Continuation-In-Part US7231097B2 (en) | 2003-12-16 | 2003-12-16 | Lens defect correction |
US11/498,427 Continuation US7259917B2 (en) | 2003-12-16 | 2006-08-02 | Image correction using a microlens array as a unit |
US11/811,356 Continuation-In-Part US7742233B2 (en) | 2003-12-16 | 2007-06-07 | Image correction using a microlens array as a unit |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/498,427 Continuation-In-Part US7259917B2 (en) | 2003-12-16 | 2006-08-02 | Image correction using a microlens array as a unit |
US11/811,356 Continuation US7742233B2 (en) | 2003-12-16 | 2007-06-07 | Image correction using a microlens array as a unit |
Publications (1)
Publication Number | Publication Date |
---|---|
US20070280550A1 true US20070280550A1 (en) | 2007-12-06 |
Family
ID=34654243
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/738,626 Active 2025-04-19 US7231097B2 (en) | 2003-12-16 | 2003-12-16 | Lens defect correction |
US11/804,314 Abandoned US20070280550A1 (en) | 2003-12-16 | 2007-05-15 | Lens defect correction |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/738,626 Active 2025-04-19 US7231097B2 (en) | 2003-12-16 | 2003-12-16 | Lens defect correction |
Country Status (5)
Country | Link |
---|---|
US (2) | US7231097B2 (en) |
JP (1) | JP2007514248A (en) |
CN (1) | CN100504489C (en) |
DE (1) | DE112004002476T5 (en) |
WO (1) | WO2005060600A2 (en) |
Families Citing this family (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
TW200822691A (en) * | 2006-11-01 | 2008-05-16 | Altek Corp | Aberration compensation system of digital camera |
US8897595B2 (en) * | 2008-03-26 | 2014-11-25 | Ricoh Co., Ltd. | Adaptive image acquisition for multiframe reconstruction |
US9865043B2 (en) | 2008-03-26 | 2018-01-09 | Ricoh Company, Ltd. | Adaptive image acquisition and display using multi-focal display |
US9866826B2 (en) | 2014-11-25 | 2018-01-09 | Ricoh Company, Ltd. | Content-adaptive multi-focal display |
JP5528139B2 (en) * | 2010-02-02 | 2014-06-25 | キヤノン株式会社 | Image processing apparatus, imaging apparatus, and image processing program |
JP5818514B2 (en) * | 2011-05-27 | 2015-11-18 | キヤノン株式会社 | Image processing apparatus, image processing method, and program |
US9864205B2 (en) | 2014-11-25 | 2018-01-09 | Ricoh Company, Ltd. | Multifocal display |
JP6924761B2 (en) * | 2015-12-18 | 2021-08-25 | ベンタナ メディカル システムズ, インコーポレイテッド | Systems and methods for separating images with different acquisition properties |
CN112308832B (en) * | 2020-10-29 | 2022-07-22 | 常熟理工学院 | Bearing quality detection method based on machine vision |
Citations (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4141032A (en) * | 1976-12-08 | 1979-02-20 | Ernst Leitz Wetzlar Gmbh | Method of and apparatus for the expansion of the range of the depth of focus beyond the limit given by conventional images |
US4275949A (en) * | 1977-04-13 | 1981-06-30 | Jones Neal K | Method of and means for scanning images |
US4584704A (en) * | 1984-03-01 | 1986-04-22 | Bran Ferren | Spatial imaging system |
US5451766A (en) * | 1993-07-12 | 1995-09-19 | U.S. Philips Corporation | Imaging device containing an electrooptic material to adjust the effective focal length of a lens element |
US5581637A (en) * | 1994-12-09 | 1996-12-03 | Xerox Corporation | System for registering component image tiles in a camera-based scanner device transcribing scene images |
US5751863A (en) * | 1996-06-18 | 1998-05-12 | Hewlett Packard Company | Method and system having relaxed front end distortion requirements |
US5805348A (en) * | 1996-08-12 | 1998-09-08 | Eastman Kodak Company | Lens system with reduced sensitivities and a method of manufacturing |
US5943050A (en) * | 1994-04-07 | 1999-08-24 | International Business Machines Corporation | Digital image capture control |
US6088083A (en) * | 1997-10-22 | 2000-07-11 | Carl-Zeiss-Stiftung | Optical image recording arrangement and method of using the same |
US6115146A (en) * | 1996-06-19 | 2000-09-05 | Asahi Kogaku Kogyo Kabushiki Kaisha | Scanning-type image reading device |
US20020176011A1 (en) * | 2001-05-22 | 2002-11-28 | Fuji Photo Film Co., Ltd. | On-screen device for subject of interest in portable electronic device, and method of controlling same |
US20030122936A1 (en) * | 2002-01-02 | 2003-07-03 | Pyle Norman C. | Proxy-based remote control method and system for a digital image capture device |
US6967780B2 (en) * | 2004-01-21 | 2005-11-22 | Searete Llc | Image correction using individual manipulation of microlenses in a microlens array |
US7012757B2 (en) * | 2002-05-08 | 2006-03-14 | Be Here Corporation | Method for designing a lens system and resulting apparatus |
US20060098861A1 (en) * | 2002-07-18 | 2006-05-11 | See Chung W | Image analysis method, apparatus and software |
US7046401B2 (en) * | 2001-06-01 | 2006-05-16 | Hewlett-Packard Development Company, L.P. | Camera-based document scanning system using multiple-pass mosaicking |
US7058233B2 (en) * | 2001-05-30 | 2006-06-06 | Mitutoyo Corporation | Systems and methods for constructing an image having an extended depth of field |
US7308158B2 (en) * | 2003-06-20 | 2007-12-11 | Eastman Kodak Company | Imaging method and system |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH06311411A (en) * | 1993-03-15 | 1994-11-04 | Toshiba Corp | Image processor |
JPH10290389A (en) * | 1997-04-16 | 1998-10-27 | Toyota Motor Corp | Multi-focus image formation method and image formation device |
JP4136044B2 (en) * | 1997-12-24 | 2008-08-20 | オリンパス株式会社 | Image processing apparatus and image processing method therefor |
JP4097761B2 (en) * | 1998-03-02 | 2008-06-11 | オリンパス株式会社 | Autofocus microscope and autofocus detection device |
JP3956584B2 (en) * | 2000-06-02 | 2007-08-08 | 株式会社日立製作所 | Omnifocal image composition method and apparatus |
US20030098352A1 (en) * | 2000-11-24 | 2003-05-29 | Metrologic Instruments, Inc. | Handheld imaging device employing planar light illumination and linear imaging with image-based velocity detection and aspect ratio compensation |
US20020114077A1 (en) * | 2001-01-23 | 2002-08-22 | Bahram Javidi | Integral three-dimensional imaging with digital reconstruction |
JP2002351439A (en) * | 2001-05-25 | 2002-12-06 | Fuji Photo Film Co Ltd | Digital camera with on-screen function and its operation control method |
US6631991B2 (en) * | 2001-08-31 | 2003-10-14 | Adaptive Optics Associates, Inc. | Ophthalmic instrument having hartmann wavefront sensor deriving location of spots with spot fitting techniques |
US7251078B2 (en) * | 2004-01-21 | 2007-07-31 | Searete, Llc | Image correction using a microlens array as a unit |
-
2003
- 2003-12-16 US US10/738,626 patent/US7231097B2/en active Active
-
2004
- 2004-12-08 DE DE112004002476T patent/DE112004002476T5/en not_active Withdrawn
- 2004-12-08 JP JP2006545737A patent/JP2007514248A/en active Pending
- 2004-12-08 WO PCT/US2004/041038 patent/WO2005060600A2/en active Application Filing
- 2004-12-08 CN CNB2004800402714A patent/CN100504489C/en not_active Expired - Fee Related
-
2007
- 2007-05-15 US US11/804,314 patent/US20070280550A1/en not_active Abandoned
Patent Citations (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4141032A (en) * | 1976-12-08 | 1979-02-20 | Ernst Leitz Wetzlar Gmbh | Method of and apparatus for the expansion of the range of the depth of focus beyond the limit given by conventional images |
US4275949A (en) * | 1977-04-13 | 1981-06-30 | Jones Neal K | Method of and means for scanning images |
US4584704A (en) * | 1984-03-01 | 1986-04-22 | Bran Ferren | Spatial imaging system |
US5451766A (en) * | 1993-07-12 | 1995-09-19 | U.S. Philips Corporation | Imaging device containing an electrooptic material to adjust the effective focal length of a lens element |
US5943050A (en) * | 1994-04-07 | 1999-08-24 | International Business Machines Corporation | Digital image capture control |
US5581637A (en) * | 1994-12-09 | 1996-12-03 | Xerox Corporation | System for registering component image tiles in a camera-based scanner device transcribing scene images |
US5751863A (en) * | 1996-06-18 | 1998-05-12 | Hewlett Packard Company | Method and system having relaxed front end distortion requirements |
US6115146A (en) * | 1996-06-19 | 2000-09-05 | Asahi Kogaku Kogyo Kabushiki Kaisha | Scanning-type image reading device |
US5805348A (en) * | 1996-08-12 | 1998-09-08 | Eastman Kodak Company | Lens system with reduced sensitivities and a method of manufacturing |
US6088083A (en) * | 1997-10-22 | 2000-07-11 | Carl-Zeiss-Stiftung | Optical image recording arrangement and method of using the same |
US20020176011A1 (en) * | 2001-05-22 | 2002-11-28 | Fuji Photo Film Co., Ltd. | On-screen device for subject of interest in portable electronic device, and method of controlling same |
US7058233B2 (en) * | 2001-05-30 | 2006-06-06 | Mitutoyo Corporation | Systems and methods for constructing an image having an extended depth of field |
US7046401B2 (en) * | 2001-06-01 | 2006-05-16 | Hewlett-Packard Development Company, L.P. | Camera-based document scanning system using multiple-pass mosaicking |
US20030122936A1 (en) * | 2002-01-02 | 2003-07-03 | Pyle Norman C. | Proxy-based remote control method and system for a digital image capture device |
US7012757B2 (en) * | 2002-05-08 | 2006-03-14 | Be Here Corporation | Method for designing a lens system and resulting apparatus |
US20060098861A1 (en) * | 2002-07-18 | 2006-05-11 | See Chung W | Image analysis method, apparatus and software |
US7308158B2 (en) * | 2003-06-20 | 2007-12-11 | Eastman Kodak Company | Imaging method and system |
US6967780B2 (en) * | 2004-01-21 | 2005-11-22 | Searete Llc | Image correction using individual manipulation of microlenses in a microlens array |
Also Published As
Publication number | Publication date |
---|---|
CN100504489C (en) | 2009-06-24 |
DE112004002476T5 (en) | 2006-10-26 |
WO2005060600A3 (en) | 2006-11-02 |
US7231097B2 (en) | 2007-06-12 |
US20050129327A1 (en) | 2005-06-16 |
WO2005060600A2 (en) | 2005-07-07 |
JP2007514248A (en) | 2007-05-31 |
CN1942802A (en) | 2007-04-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20070280550A1 (en) | Lens defect correction | |
US7399082B2 (en) | Volumetric imaging using “virtual” lenslets | |
US7259917B2 (en) | Image correction using a microlens array as a unit | |
US9348123B2 (en) | Image correction using individual manipulation of microlenses in a microlens array | |
US6967780B2 (en) | Image correction using individual manipulation of microlenses in a microlens array | |
US9288395B2 (en) | Super-resolution based on optical image stabilization | |
EP1999947B1 (en) | Image capturing device with improved image quality | |
JP2008245157A (en) | Imaging device and method therefor | |
US7826139B2 (en) | Image correction using individual manipulation of microlenses in a microlens array | |
KR20210137417A (en) | Optical system and camera module for comprising the same | |
KR20200081057A (en) | Method and Apparatus for Center Calibration of Camera System | |
US7417797B2 (en) | Image correction using individual manipulation of microlenses in a microlens array | |
US7742233B2 (en) | Image correction using a microlens array as a unit | |
WO2005081836A2 (en) | Defect correction based on 'virtual' lenslets | |
US20070025640A1 (en) | Control and/or making aspects related to chromatic aberration correction by moving photo-detector array | |
US11463674B1 (en) | Imaging system and display apparatus incorporating super resolution using fixed focus cameras | |
US7840038B2 (en) | Defect correction based on “virtual” lenslets | |
US8169534B2 (en) | Volumetric imaging using “virtual” lenslets | |
WO2005070149A2 (en) | Image correction using a microlens array as a unit | |
US8243148B2 (en) | Image processing device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SEARETE LLC, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:HILLIS, W. DANIEL;MYHRVOLD, NATHAN P.;WOOD, LOWELL L. WOOD JR.;REEL/FRAME:019727/0945;SIGNING DATES FROM 20070622 TO 20070626 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |