US9121702B2 - Distance measurement apparatus and distance measurement method - Google Patents

Distance measurement apparatus and distance measurement method Download PDF

Info

Publication number
US9121702B2
US9121702B2 US13/521,109 US201113521109A US9121702B2 US 9121702 B2 US9121702 B2 US 9121702B2 US 201113521109 A US201113521109 A US 201113521109A US 9121702 B2 US9121702 B2 US 9121702B2
Authority
US
United States
Prior art keywords
distance
subject
image
distance measurement
diffractive optical
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US13/521,109
Other languages
English (en)
Other versions
US20120281086A1 (en
Inventor
Masayuki Kimura
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Panasonic Corp
Original Assignee
Panasonic Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Panasonic Corp filed Critical Panasonic Corp
Assigned to PANASONIC CORPORATION reassignment PANASONIC CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: KIMURA, MASAYUKI
Publication of US20120281086A1 publication Critical patent/US20120281086A1/en
Application granted granted Critical
Publication of US9121702B2 publication Critical patent/US9121702B2/en
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C3/00Measuring distances in line of sight; Optical rangefinders
    • G01C3/02Details
    • G01C3/06Use of electric means to obtain final indication
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S11/00Systems for determining distance or velocity not using reflection or reradiation
    • G01S11/12Systems for determining distance or velocity not using reflection or reradiation using electromagnetic waves other than radio waves
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/67Focus control based on electronic image sensor signals
    • H04N5/23212

Definitions

  • the present invention relates to a distance measurement apparatus which measures a depth of a scene based on a plurality of images captured from a single view point.
  • Various methods have been suggested for measuring, without contact, a depth of a three-dimensional scene, that is, a distance to each subject.
  • Such methods can be classified into an active method and a passive method.
  • a subject is irradiated with infrared rays, ultrasonic waves, or laser beams so as to calculate the subject based on a length of time until a wave which is reflected returns or an angle of the reflected wave.
  • the passive method the distance is calculated based on an image of the subject.
  • the passive method which does not require an apparatus for emitting infrared rays and so on is widely used.
  • DFD Depth from Defocus
  • h denotes a Point Spread Function (hereinafter, referred to as PSF) which represents a blur condition in an optical system
  • d denotes the distance to the subject in the position (x, y) on the captured image or the original image.
  • PSF Point Spread Function
  • d denotes the distance to the subject in the position (x, y) on the captured image or the original image.
  • the h represents a function which depends on the aforementioned position (x, y) and the subject distance d.
  • * in the expression represents convolution operation.
  • the expression 1 includes S and d as unknown quantities.
  • images I 2 for a single scene are captured, changing focal positions. Change in the focal position corresponds to change in the PSF with respect for a single subject distance.
  • h′ denotes a PSF in a focal position different from the focal position for h.
  • the original image S and the subject distance d in the scene can be obtained by solving these expressions.
  • Various solutions for the original image S and the subject distance d have been suggested in Patent Literature 1 and others.
  • Patent Literature 1 Japanese Patent Gazette 2963990
  • Patent Literature 2 Japanese Unexamined Patent Application Publication No. 62-222117
  • Patent Literature 3 Japanese Unexamined Patent Application Publication No. 2006-329897
  • the present invention is made to solve the above problems, and an object of the present invention is to provide a distance measurement apparatus and distance measurement method with high distance-measurement accuracy even in a condition where a depth of field is large, such as a short focal distance.
  • a distance measurement apparatus includes: an imaging device which captures an image; a diffractive optical device which diffracts a subject image; an optical system which forms, on the imaging device, the image from the subject image diffracted by the diffractive optical device; and a distance measurement unit which measures a distance to a subject using the image captured by the imaging device, in which the distance measurement unit measures the distance to the subject based on an interval between diffraction figures of the subject image which are created by the diffractive optical device, and the interval is on the image captured by said imaging device.
  • the distance to the subject is calculated based on the interval between the diffraction figures which are created by the diffraction in addition to a size of a blur.
  • the interval between the diffraction figures are not affected by a depth of field. The shorter the distance to the subject is, the narrower the interval becomes, i.e., the longer the distance to the subject is, the wider the interval becomes. Accordingly, even if the depth of field is large and the blur does not occur, the distance to the subject can accurately be calculated using images captured from a single view point.
  • the aforementioned distance measurement apparatus may further include an optical filter which transmits each of light beams in a corresponding one of specified wavelength bands.
  • the effects due to the diffraction vary according to wavelengths. If no optical filter is used, a light beam from a single point receives various effects due to the diffraction. Accordingly, the light beam is not focused on a single point, causing the image to be blurred. This requires processing of removing the blur. According to the configuration of the present invention, since only a specified wavelength is used to obtain an image having no blur, no processing of removing the blur is required. This simplifies processing of calculating the distance to the subject.
  • the diffractive optical device may have a curved surface.
  • the curved surface is preferably a spherical surface, and a curvature radius of the curved surface is approximately equal to a distance between the aperture of the optical system and the diffractive optical device on the optical axis.
  • the curved surface may be an aspheric surface.
  • the aforementioned distance measurement apparatus may further include a lighting device which emits a light beam in a specified wavelength band to the subject.
  • the light beam in the specified wavelength band is emitted to the subject, causing the image in the specified wavelength band to be captured, in the same manner with the case where the optical filter is used. Accordingly, an image having no blur can be obtained, and the processing of calculating the distance to the subject can be simplified.
  • the present invention can be achieved not only as a distance measurement apparatus including such distinguishable processing units but also as a distance measurement method including, as steps, processing implemented by the distinguishable processing units included in the distance measurement apparatus.
  • the present invention can also be achieved as a program which causes a computer to be operated as the distinguishable processing units included in the distance measurement apparatus, or a program which causes a computer to execute the distinguishable steps included in the distance measurement method
  • a program can be distributed via a non-transitory computer readable recording medium including Compact Disc-Read Only Memory (CD-ROM) and so on or a communication network such as the Internet.
  • CD-ROM Compact Disc-Read Only Memory
  • a distance to a subject is calculated based on an interval between diffraction figures generated not by a size of a blur but by diffraction. Accordingly, the distance to the subject can be accurately obtained from images captured at a single view point even in a condition where a depth of field is large and the blur does not occur.
  • FIG. 1 is a block diagram showing a functional construction of a distance measurement apparatus according to Embodiment 1 of the present invention.
  • FIG. 2 is a diagram schematically showing incidence of a light beam via the diffractive optical device.
  • FIG. 3 is a diagram showing a geometric relationship on the incidence of the light beam via the diffractive optical device.
  • FIG. 4 is a flow chart showing an example of a flow of processing of calculating a subject distance.
  • FIG. 5 is a diagram for illustrating movement of the diffractive optical device by an actuator in order to capture a diffracted image.
  • FIG. 6 is a diagram for illustrating movement of the diffractive optical device by the actuator in order to capture a reference image.
  • FIG. 7A is a diagram schematically showing a PSF shape (kernel shape) in the case of a short subject distance.
  • FIG. 7B is a diagram schematically showing a PSF shape (kernel shape) in the case of a long subject distance.
  • FIG. 8 is a diagram schematically showing the incidence of the light beam from a point off an optical axis through the diffractive optical device.
  • FIG. 9A is a diagram schematically showing the PSF shape (kernel shape) in the case of the short subject distance.
  • FIG. 9B is a diagram schematically showing the PSF shape (kernel shape) off the optical axis.
  • FIG. 10 is a diagram schematically showing the incidence of the light beam from a point off the optical axis via the diffractive optical device formed in a spherical surface.
  • FIG. 11 is a diagram showing an example of a diffracted image obtained by capturing a white-point light source without using an optical filter.
  • FIG. 12A is a diagram showing an example of the diffracted image obtained by capturing the white-point light source using an optical filter in a blue wavelength band.
  • FIG. 12B is a diagram showing an example of the diffracted image obtained by capturing the white-point light source using an optical filter in a green wavelength band.
  • FIG. 12C is a diagram showing an example of the diffracted image obtained by capturing the white-point light source using an optical filter in a red wavelength band.
  • FIG. 13 is a block diagram showing a functional construction of the distance measurement apparatus including components necessary for the present invention.
  • FIG. 1 is a block diagram showing a construction of a distance measurement apparatus in view of its function according to Embodiment 1 of the present invention.
  • a distance measurement apparatus 10 includes an imaging unit 12 , a frame memory 14 , a distance measurement unit 16 , and a control unit 18 .
  • the imaging unit 12 includes an imaging device 20 , a diffractive optical device 24 , an optical system 22 , an optical filter 26 , and an actuator 28 .
  • the imaging unit 12 captures a subject image and provides an image.
  • the imaging device 20 captures an image, and includes a Charge Coupled Device (CCD), a Complementary Metal Oxide Semiconductor (CMOS), or the like.
  • CCD Charge Coupled Device
  • CMOS Complementary Metal Oxide Semiconductor
  • the diffractive optical device 24 causes the subject image to be diffracted.
  • a preferable material of the diffractive optical device 24 includes, but not particularly limited to, a transparent glass or the like which transmits light. Transmission of light enables to prevent light quantity from decreasing.
  • the diffractive optical device 24 is constructed such that parallel grooves are formed in a glass to form a convexo-concave surface or transparent materials having refractive indexes different from each other are alternately arranged. Meanwhile, the diffractive optical device 24 can also be prepared by alternately arranging a portion which blocks light and a portion which transmits light. In this case, however, the light quantity decreases.
  • the optical system 22 forms, in the imaging device 20 , an image of the subject image diffracted by the diffractive optical device 24 .
  • the optical filter 26 transmits a light beam having a specified wavelength band.
  • the actuator 28 allows the diffractive optical device 24 to advance or retreat with respect to an optical path of the optical system 22 .
  • the diffractive optical device 24 is movable back and forth with respect to the optical path by the actuator 28 , so as to switchingly capture images between a image in which diffraction does not occur and an image in which the diffraction occurs.
  • the diffractive optical device 24 is located in a position which is on an optical axis of the optical system 22 and is closer to a subject than an aperture of the optical system 22 is.
  • the frame memory 14 is a memory for storing the image on a basis of frame, and stores the image provided from the imaging unit 12 , and so on.
  • the distance measurement unit 16 calculates a subject distance based on the image captured by the imaging unit 12 . Specifically, the distance measurement unit 16 measures the distance to the subject based on intervals of diffraction figures of the subject image on the image captured by the imaging device 20 . The diffraction figures are created by the diffractive optical device 24 .
  • the control unit 18 includes a central processing unit (CPU), as well as a read only memory (ROM), a random access memory (RAM) and others for storing a control program, so as to control functional blocks.
  • CPU central processing unit
  • ROM read only memory
  • RAM random access memory
  • FIG. 2 is a diagram schematically showing that a light beam is incident via the diffractive optical device.
  • a diffractive optical device G corresponds to the diffractive optical device 24
  • an imaging surface I corresponds to the imaging device 20 .
  • the diffractive optical device G is located between a subject o and a pinhole P. The light beam emitted from the subject o enters the diffractive optical device G, and is inflected at a diffraction angle in response to a property of the diffractive optical device G. Among the inflected light beams, only a light beam which has a specified inflection angle can pass the pinhole P, and reach the imaging surface I.
  • the diffraction angle is defined in accordance with an order of the diffraction. Accordingly, a plurality of light beams including a zero-order light beam (L 0 ) which is not a diffracted light beam, a positive first-order diffracted light beam (L 1 ), a negative first-order diffracted light beam (L- 1 ), and so on reach the imaging surface I. Although a diffracted light beam of second order or higher can reach the imaging surface I, those are not shown in FIG. 2 for simplification. It should be noted that a position of the pinhole P corresponds to a position of the aperture of the optical system 22 .
  • FIG. 3 is a diagram showing a geometric aspect in a case when a light beam is incident via the diffractive optical device 24 (G).
  • f denotes a distance from the pinhole to the imaging surface, i.e., a focal distance.
  • an angle ⁇ is equal to an angle of emergence of the light beam from the diffractive optical device G. Accordingly, if an incident angle of the light beam from the subject o to the diffraction optical element G is set to ⁇ , Expression 4 shown below is satisfied based on the formula of diffraction.
  • n, ⁇ , and D respectively represents the order of the diffraction, a wavelength of the incident light, and a lattice spacing of the diffractive optical device G.
  • Expression 5 shown below can be established between a distance v which is defined from the subject o to the diffractive optical device G on the optical axis and a distance u which is defined from the diffractive optical device G to the pinhole P on the optical axis.
  • v tan ⁇ u tan ⁇
  • the angle ⁇ associated with the order n of the diffraction is calculated from Expressions 4 and 5, and the distance d can be calculated from Expression 3 based on the obtained angle ⁇ . Furthermore, the angle ⁇ is determined in accordance with the distance v to the subject, in view of Expression 5. Accordingly, a position d (which is determined by the distance d from the center of the optical axis on the imaging surface I and at which the diffraction figure is created) also changes in accordance with the subject distance.
  • Patent Literature 1 discloses a method for measuring a distance to a subject using the diffractive optical device, similar to the present invention.
  • Patent Literature 2 discloses a method for calculating a distance to a subject based on intervals of light points of light beams when the light beams reach the subject, the light beams being dispersed by the diffraction which is generated by irradiating a transmissive diffractive optical device with a laser beam. According to the method, the subject distance is calculated using the principle of triangulation.
  • FIG. 4 is a flow chart showing an example of the flow of processing in which the distance measurement apparatus 10 calculates the subject distance. According to this processing, the most appropriate subject distance is calculated from predetermined n stages of subject distances V( 1 ), V( 2 ), . . . , and V(n).
  • the imaging device 20 captures a diffracted image I and a reference image I′ (Steps S 101 and S 102 ).
  • the diffracted image represents an image obtained in a condition that the diffractive optical device 24 is placed on the optical path
  • the reference image represents an image obtained in a condition that the diffractive optical device 24 is not placed on the optical path.
  • the actuator 28 causes the diffractive optical device 24 to advance on the optical path, as shown in FIG. 5 .
  • the actuator 28 causes the diffractive optical device 24 to retreat from the optical path, as shown in FIG. 6 .
  • the Steps S 101 and S 102 may be in the reverse order.
  • FIG. 7A or FIG. 7B denotes a kernel schematically shown in FIG. 7A or FIG. 7B .
  • h denotes a kernel schematically shown in FIG. 7A or FIG. 7B .
  • FIGS. 7A and 7B each show examples of the kernel when three light beams are considered, as the diffracted light beams, including the zero-order light beam, the positive first-order light beam, and the negative first-order light beam.
  • FIGS. 7A and 7B each show examples using, as the optical system, an optical system in which a degree of the blur undergoes little change according to the distance.
  • the kernel can be calculated by simulation, and corresponds to the ideal diffracted image I obtained when a point light source is assumed to exist at a position of the subject, for example.
  • the control unit 18 substitutes an initial value 1 into a counter i (Step S 103 ).
  • the distance measurement unit 16 calculates a value of an error function C(x, y, i) with respect to the subject distance V(i) at the stage I, for every pixel (x, y) of an image (the diffracted image I or the reference image I′) (Step S 104 ).
  • the value of the error function C(x, y, i) corresponds to a difference between an actual diffracted image I(x, y) and a value resulting from the convolution operation on the reference image I′(x, y) which is not subjected to the diffraction with the kernel h(x, y, i) corresponding to the distance V(i) at the stage i.
  • the difference is minimum when the object o actually exist at the distance V(i) in the stage i.
  • the error function C(x, y, i) expresses an absolute value of the difference between pixels in Expression 7.
  • the error function C(x, y, i) can be defined based on an arbitrary form which expresses a distance, such as L 2 norm.
  • the control unit 18 calculates the value of the error function C(x, y, i), and then determines whether or not a value of the counter i reaches n (Step S 105 ). If the value of the counter i does not reach n (NO in Step S 105 ), the control unit 18 increments the value of the counter i by 1 (Step S 106 ). The distance measurement unit 16 repeats processing of calculating the value of the error function C(x, y, i) (Step S 104 ) until the value of the counter i reaches n.
  • the distance measurement unit 16 calculates all values of the error function C(x, y, i) in the stage n (YES, in Step S 105 ), and subsequently, calculates the subject distance (Step S 107 ).
  • the subject distance v(x, y) at the position (x, y) can be represented by Expression 8 shown below.
  • Expression 9 represents a value of i when a value of a function f(i) becomes minimum.
  • the distance V(i) associated with the stage i where the value of the error function C(x, y, i) is minimum is calculated as the subject distance at each position (x, y).
  • a value of a constant a is preferably determined as a size such that the diffracted image in the position d(i) of each stage i is distinguishable on the captured image (for example, a level of a pixel size of the imaging device).
  • the subject distance V(i) may be determined in accordance with relationships other than the above.
  • Expressions 3 to 5 are determined without considering effects of refraction or the like due to a thickness or a material of the diffractive optical device 24 . Accordingly, the position d(i) of the diffraction figure, which takes into account these effects, may be calculated by an optical simulation to determine the corresponding subject distance V(i), for example.
  • Expression 6 is expressed by the same form with those of Expressions 1 and 2. Accordingly, the blurs in the diffraction figure and DFD can be expressed in a same frame. Therefore, processing of DFD may be applied to calculate the subject distance in the recognition of the kernel indicated in FIG. 7A or 7 B as the PSF. In this case, an existing DFD processing algorithm is used to calculate the subject distance.
  • each of the intervals of the diffraction figures may directly be calculated by autocorrelation or the like from the diffraction figure I, so as to calculate the subject distance using Expressions 3 to 5.
  • Patent Literature 3 discloses a method, for example, for calculating an interval of double images shown on a transparent plate by searching a position in which the autocorrelation takes the local maximal value along displacement constraint between the double images. Such a method may be used to calculate the subject distance.
  • the subject distance can be calculated based on the interval between the diffraction figures, while the conventional DFD calculates the subject distance based on variation in the size of the blur.
  • the variation amount is approximately proportional to a size of the aperture of the optical system. Accordingly, the variation amount becomes small as the focal distance decreases.
  • the interval d between the diffraction figures is determined based on the distance u between the diffractive optical device G and the pinhole P, a lattice spacing D, and the order n of the diffraction, but not based on a focal distance f (the distance from the pinhole to the imaging surface) and the size of the aperture. Accordingly, even if the focal distance is short, the variation amount of the interval d between the diffraction figures can be increased. As a result, measurement accuracy of the subject distance can be enhanced.
  • a distance measurement apparatus can keep intervals between diffraction figures equal to each other with respect to a diffraction direction.
  • the diffractive optical device 24 has a curved-surface shape.
  • Other components are the same with those in Embodiment 1. The reason for the construction is described below.
  • FIGS. 2 and 3 the case where the subject o is on an optical axis is considered.
  • FIG. 8 the case where the subject o is on a position defined by a field angle ⁇ is considered.
  • a light beam entering the diffractive optical device G( 24 ) at a specified angle among light beams emitted from the subject o is inflected by the diffraction, subsequently passes through the pinhole P, and then enters the imaging surface I.
  • a value d′ can be calculated from Expressions 10 and 11, and Expression 4 which is the formula of the diffraction.
  • R, d, nd, vd respectively represent, in this order, a curvature radius (unit: mm) of each surface, a surface interval (unit: mm), a refractive index of d line, and an Abbe number.
  • the surface number 0 indicates a position of the object.
  • a surface having a surface type of “diffraction” indicates a diffraction surface
  • a surface having the surface type of “paraxial” indicates an ideal lens with no aberration.
  • a surface having a blank for the surface type indicates a spherical surface.
  • a lattice spacing of the diffraction surface is set to 2 ⁇ m
  • a focal distance of the ideal lens is set to 5 nm
  • a wavelength of the light beam is set to 550 nm.
  • each value of d′(+1), d′(0), and d′( ⁇ 1) is calculated for the respective cases when the field angle ⁇ is 0 degree, 10 degrees, and 20 degrees.
  • the obtained values are shown in Table 2. Every unit for d′ is mm.
  • xey indicates x ⁇ 10 y .
  • the positive first-order light beam and the negative first-order light beam are located in respective positions each equally distant away from the zero-order light beam, as schematically shown in FIG. 9A .
  • the intervals of the diffracted light beams suffer from non-uniformity, as shown in FIG. 9B .
  • the kernel need to be prepared for every position (x, y) in images even in the same ith distance, as the kernel h(x, y, i) which is the subject to the convolution operation when the value of the error function C(x, y, i) is calculated in Expression 7.
  • the kernel is preferably uniform in the entire image.
  • the diffractive optical device G is formed in a curved-surface shape, and its curvature radius is defined to be approximately equal to a distance from a diffraction surface to an aperture on the optical axis.
  • the light beam traveling to the center of the pinhole P is always incident to the diffractive optical device G approximately perpendicularly. Therefore, the incident angle can be kept approximately constant.
  • a position of the diffraction figure at the time when the diffraction surface is in the curved-surface shape is calculated as follows.
  • the curvature radius of the diffraction surface is equal to the distance from the diffraction surface to the aperture, so that an angle of emergence after the diffraction becomes orthogonal to the diffraction surface. Therefore, Expression 13 is satisfied from the formula of diffraction.
  • n, ⁇ , and D respectively denotes an order of the diffraction, a wavelength of the incident light beam, and a lattice spacing of the diffractive optical device G. Deletion of a from Expressions 12 and 13 allows an angle ⁇ associated with the field angle ⁇ to be calculated, thereby obtaining a position d of the diffraction figure from Expression 3.
  • a flow of processing of calculating the subject distance when the diffraction surface is the curved-surface shape is approximately same with those described in Embodiment 1.
  • the difference from Embodiment 1 is that the subject distance is calculated in accordance with Expression 14 shown below instead of Expression 8, from the subject distance V(x, y, 1), V(x, y, 2), . . . , and V(x, y, n) at the stage n previously determined for each of pixels in the image.
  • V uses a single sequence in the overall image, while in Expression 14, a used sequence varies according to each of pixels.
  • each of the intervals between the diffraction figures varies depending on the field angle in addition to the positive and negative of the order of the diffraction.
  • the diffraction surface is formed in the curved-surface shape, allowing a light beam to enter the diffraction surface perpendicularly. This solves the non-uniformity occurring in the intervals between the diffraction figures. Therefore, the subject distance calculation processing can be simplified.
  • the diffractive optical device 24 has an aspheric curved-surface shape.
  • Use of the aspheric-surface shape makes it possible to control, in detail, a position where a light beam reaches, thereby keeping intervals between diffraction figures to be equal to each other with respect to positive and negative of an order of diffraction, and allowing the interval to be constant with respect to a field angle.
  • a specific shape of the aspheric surface is determined by performing optimization by an optical simulation in such a manner that, in an arbitrary field angle and wavelength, the intervals of the diffraction figures are equal to each other with respect to the positive and negative of the order of the diffraction, and are constant with respect to the field angle.
  • a lattice spacing of a diffraction surface, a focal distance of an ideal lens, and a wavelength of the light beam are respectively set to 2 ⁇ m, 5 mm, and 550 nm.
  • the field angle ⁇ is estimated with respect to three points, i.e., 0 degree, 10 degrees, and 20 degrees.
  • a surface having a surface type of “aspherical” represents an aspheric surface.
  • the aspheric-surface shape is represented by Expression 15 shown below.
  • k denotes a constant of the cone
  • A4, A6, A8, A10, and A12 respectively denote fourth, sixth, eighth, tenth, and twelfth aspheric coefficients.
  • the constant of the cone and the aspheric coefficient with respect to the second surface in Table 5 are shown in Table 6.
  • r denotes a curvature of the aspheric shape
  • z denotes a distance in the optical axis direction from an aperture to the aspheric-surface shape.
  • Table 7 shows respective values of d′(+1), d′(0), and d′( ⁇ 1) respectively obtained when the field angle ⁇ is set to 0 degree, 10 degrees, and 20 degrees in the optical system indicated in Tables 5 and 6.
  • Table 8 shows a position of the diffraction figure when the subject distance is set to 1,000 mm in the same optical system.
  • the present invention is not limited to the above, but applicable to various embodiments for achieving at an object of the present invention and other objects associated therewith.
  • the present invention may also include the following.
  • a pinhole is used as the optical system for calculating the position of the diffracted image in Embodiments 1 and 2.
  • an optical imaging system may be used which includes a single lens or a lens group consisting of a plurality of lenses.
  • the positions of the diffraction figures are different from those in the case when the pinhole is used, due to a refraction effect of the lens.
  • the positions of the diffraction figures are also changed depending on the subject distance. Accordingly, the subject distance can be calculated in the same manner with the case when the pinhole is used. In other words, the subject distance can be calculated in the same manner with the case when the pinhole is used, by assuming that the position of the aperture of the optical system corresponds to the position of the pinhole.
  • the optical filter 26 may have a feature of allowing light beams in a plurality of wavelength bands to pass therethrough, and the subject distance may be calculated from images in the wavelength bands which were allowed to pass through the optical filter.
  • FIG. 11 is a diagram showing an example of a diffracted image obtained by capturing a white-point light source without using the aforementioned optical filter 26 .
  • Expression 4 shows that the angle of emergence of the diffracted light beam varies depending on the wavelength band. For this reason, the diffraction figures are blurred in FIG. 11 .
  • FIGS. 12A to 12C show the diffracted images in which the white-point light source which is the same with that used in FIG. 11 is captured in each of wavelength bands, using the aforementioned optical filter 26 .
  • FIGS. 12A , 12 B, and 12 C respectively show the diffracted images in a blue wavelength band, a green wavelength band, and a red wavelength band. No blur occurs in the diffraction figures in the respective wavelength bands.
  • the distance-measurement results obtained from the respective diffracted images are combined, so that an error in the subject-distance calculation due to noise in the image is reduced, thereby enabling to measure the distance with higher accuracy.
  • an average of a plurality of distances at a single pixel position, which is measured using a plurality of diffracted images, may be the subject distance.
  • a color filter may be used which is arranged in a front surface of the optical system.
  • the color filter is used, so as to transmit a light beam in a specified wavelength.
  • three color filters which respectively transmit light beams in the red, blue, and green wavelength bands are switchingly used to measure distances using the light beams in the respective three wavelength bands.
  • a lighting device which emits a light beam having a specified wavelength band may be used instead of the color filter.
  • Such a lighting device may emit an illumination light beam in a specified wavelength band to a subject, so that the diffracted image in a specified wavelength band may be captured.
  • the diffraction image may be captured by emitting the illumination light beam to the subject with the wavelength being switched.
  • a part or all of components constituting the aforementioned respective devices may be formed, specifically, as a computer system including a microprocessor, a ROM, a RAM, a hard disc unit, a display unit, a keyboard, a mouse, and the like.
  • the RAM or the hard disc unit stores a computer program.
  • the microprocessor operates in accordance with the computer program, so that each of the devices accomplishes its function.
  • the computer program is, for accomplishing a predetermined function, configured by combining a plurality of instruction codes indicating instructions for a computer.
  • a part or all of the components constituting each of the above devices may be configured by a single System-LSI (Large-Scale Integration) circuit.
  • the System-LSI is a super multifunction LSI manufactured by integrating a plurality of constituent units on a single chip, and is, specifically, a computer system including a microprocessor, a ROM, a RAM, and so on.
  • the RAM stores a computer program.
  • the microprocessor operates in accordance with the computer program, so that the System-LSI accomplishes its function.
  • a part or all of the components constituting each of the devices may be configured as an IC card which is detachable from each of the devices or a single module.
  • the IC card or the module is a computer system including a microprocessor, a ROM, a RAM, and so on.
  • the IC card or the module may include the super multifunction LSI.
  • the microprocessor operates in accordance with the computer program, so that the IC card or the module accomplishes its function.
  • the IC card or the module may have tamper resistance.
  • the present invention may be in the form of the method described above.
  • the present invention may be a computer program which realizes the method by a computer, or may be a digital signal including the computer program.
  • the present invention may also be realized by storing the computer program or the digital signal in a computer readable recording medium, such as a flexible disc, a hard disc, a CD-ROM, an MO, a DVD, a DVD-ROM, a DVD-RAM, a Blue-ray disc (registered mark), or a semiconductor memory.
  • a computer readable recording medium such as a flexible disc, a hard disc, a CD-ROM, an MO, a DVD, a DVD-ROM, a DVD-RAM, a Blue-ray disc (registered mark), or a semiconductor memory.
  • the present invention may also include the digital signal recorded in these recording media.
  • the present invention may also be realized by transmission of the aforementioned computer program or digital signal via an electric telecommunication line, a wireless or wired communication line, a network represented by the Internet, a data broadcast, and so on.
  • the present invention may also be a computer system including a microprocessor and a memory, in which the memory stores the aforementioned computer program, and the microprocessor operates in accordance with the computer program.
  • program or the digital signal may be stored in the recording medium so as to be transferred, or the program or the digital signal may be transferred via the network or the like so as to be executed by another independent computer system.
  • FIG. 13 is a block diagram showing a distance measurement apparatus including components necessary to the present invention.
  • a distance measurement apparatus 10 A includes an imaging unit 12 A and the distance measurement unit 16 .
  • the imaging unit 12 A includes the imaging device 20 , the optical system 22 , and the diffractive optical device 24 .
  • the distance measurement apparatus includes these components.
  • a distance measurement apparatus can measure a distance based on an image captured from a single view point. Accordingly, the distance measurement apparatus is applicable to general imaging equipment. Particularly, the distance measurement apparatus is optimal to imaging equipment using a compact imaging device requiring a focal distance to be short in view of principle of distance measurement using a position of a diffraction figure instead of blur.

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Electromagnetism (AREA)
  • General Physics & Mathematics (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Measurement Of Optical Distance (AREA)
  • Length Measuring Devices By Optical Means (AREA)
US13/521,109 2010-11-09 2011-11-07 Distance measurement apparatus and distance measurement method Expired - Fee Related US9121702B2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
JP2010250436 2010-11-09
JP2010-250436 2010-11-09
PCT/JP2011/006217 WO2012063463A1 (ja) 2010-11-09 2011-11-07 距離計測装置および距離計測方法

Publications (2)

Publication Number Publication Date
US20120281086A1 US20120281086A1 (en) 2012-11-08
US9121702B2 true US9121702B2 (en) 2015-09-01

Family

ID=46050628

Family Applications (1)

Application Number Title Priority Date Filing Date
US13/521,109 Expired - Fee Related US9121702B2 (en) 2010-11-09 2011-11-07 Distance measurement apparatus and distance measurement method

Country Status (4)

Country Link
US (1) US9121702B2 (ja)
JP (1) JP5833546B2 (ja)
CN (1) CN102695939B (ja)
WO (1) WO2012063463A1 (ja)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9572682B2 (en) 2011-09-29 2017-02-21 Arthromeda, Inc. System and method for precise prosthesis positioning in hip arthroplasty
US9597096B2 (en) 2013-03-15 2017-03-21 Arthromeda, Inc. Systems and methods for providing alignment in total knee arthroplasty

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9228829B2 (en) 2013-10-31 2016-01-05 Industrial Technology Research Institute Method and system for measuring distance
CN110220494B (zh) * 2015-10-29 2021-11-09 原相科技股份有限公司 双孔径测距系统及其运作方法
JP6549021B2 (ja) * 2015-11-17 2019-07-24 富士フイルム株式会社 内視鏡システム、及び計測装置の作動方法
CN108459417B (zh) * 2018-02-05 2020-06-26 华侨大学 一种单目窄带多光谱立体视觉系统及其使用方法
WO2019165862A1 (zh) * 2018-02-27 2019-09-06 Oppo广东移动通信有限公司 激光投射模组、深度相机和电子装置
JP7500042B2 (ja) * 2019-12-12 2024-06-17 国立大学法人 和歌山大学 光学装置、撮像システム、分析システム、方法、及び空間光変調素子の使用方法

Citations (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4310227A (en) 1981-01-05 1982-01-12 Polaroid Corporation Diffracted energy auto-ranging system for a camera
JPS62222117A (ja) 1986-03-25 1987-09-30 Mitsubishi Heavy Ind Ltd 多点距離計測センサ
WO1988005525A1 (en) 1987-01-13 1988-07-28 Thomas Dewitt Range finding by diffraction
JPH02502398A (ja) 1987-01-13 1990-08-02 デヴイット,トマス 回折による距離測定
US5557407A (en) * 1993-12-28 1996-09-17 Canon Kabushiki Kaisha Measuring apparatus having a diffraction grating for receiving a diverged light beam
WO1999044013A1 (en) 1996-12-30 1999-09-02 Thomas D Ditto Variable pitch grating for diffraction range finding system
JP2963990B1 (ja) 1998-05-25 1999-10-18 京都大学長 距離計測装置及び方法並びに画像復元装置及び方法
US20020021448A1 (en) * 2000-05-26 2002-02-21 Ko Ishizuka Measuring instrument
US20020149762A1 (en) * 2001-02-20 2002-10-17 Ditto Thomas D. Chromatic diffraction range finder
US20030189886A1 (en) 2002-04-03 2003-10-09 Konica Corporation Optical pick-up apparatus and objective lens for the optical pick-up apparatus
US20030215183A1 (en) * 2002-05-14 2003-11-20 Wildnauer Kenneth R. Tunable optical filter based on a physically-deformable diffractive element
CN1677136A (zh) 2004-04-01 2005-10-05 佳能株式会社 衍射光学元件和包括所述衍射光学元件的光学系统
US20060012772A1 (en) * 2004-07-15 2006-01-19 Mitutoyo Corporation Absolute distance measuring device
CN1844958A (zh) 2006-04-29 2006-10-11 杭州照相机械研究所 非球面衍射光学透镜
JP2006329897A (ja) 2005-05-30 2006-12-07 Tokyo Institute Of Technology 透明板に映る2重像を用いた距離計測方法
US20070086083A1 (en) * 2005-10-14 2007-04-19 Hon Hai Precision Industry Co., Ltd. Optical locating device
CN1963595A (zh) 2006-12-07 2007-05-16 四川大学 非平面光栅式三维自由立体显示屏
US20100322481A1 (en) 2009-06-19 2010-12-23 Canon Kabushiki Kaisha Image processing apparatus and method
US20110286001A1 (en) * 2009-01-26 2011-11-24 Atsushi Taniguchi Dark-field defect inspecting method, dark-field defect inspecting apparatus, aberration analyzing method, and aberration analyzing apparatus
US20130286191A1 (en) * 2010-12-27 2013-10-31 Hitachi High-Technologies Corporation Inspection apparatus

Patent Citations (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4310227A (en) 1981-01-05 1982-01-12 Polaroid Corporation Diffracted energy auto-ranging system for a camera
JPS57125913A (en) 1981-01-05 1982-08-05 Polaroid Corp Range finder for photographic camera
JPS62222117A (ja) 1986-03-25 1987-09-30 Mitsubishi Heavy Ind Ltd 多点距離計測センサ
WO1988005525A1 (en) 1987-01-13 1988-07-28 Thomas Dewitt Range finding by diffraction
JPH02502398A (ja) 1987-01-13 1990-08-02 デヴイット,トマス 回折による距離測定
US5557407A (en) * 1993-12-28 1996-09-17 Canon Kabushiki Kaisha Measuring apparatus having a diffraction grating for receiving a diverged light beam
JP2001513899A (ja) 1996-12-30 2001-09-04 ディトゥー,トーマス,ディー. 回折測距システム用可変ピッチ格子
US6490028B1 (en) 1996-12-30 2002-12-03 Thomas D. Ditto Variable pitch grating for diffraction range finding system
WO1999044013A1 (en) 1996-12-30 1999-09-02 Thomas D Ditto Variable pitch grating for diffraction range finding system
JP2963990B1 (ja) 1998-05-25 1999-10-18 京都大学長 距離計測装置及び方法並びに画像復元装置及び方法
US20020021448A1 (en) * 2000-05-26 2002-02-21 Ko Ishizuka Measuring instrument
US20020149762A1 (en) * 2001-02-20 2002-10-17 Ditto Thomas D. Chromatic diffraction range finder
US6603561B2 (en) 2001-02-20 2003-08-05 Thomas D. Ditto Chromatic diffraction range finder
US7158467B2 (en) 2002-04-03 2007-01-02 Konica Corporation Optical pick-up apparatus and objective lens for the optical pick-up apparatus
US20030189886A1 (en) 2002-04-03 2003-10-09 Konica Corporation Optical pick-up apparatus and objective lens for the optical pick-up apparatus
CN1448928A (zh) 2002-04-03 2003-10-15 柯尼卡株式会社 光拾取设备和用于该光拾取设备的物镜
US20030215183A1 (en) * 2002-05-14 2003-11-20 Wildnauer Kenneth R. Tunable optical filter based on a physically-deformable diffractive element
CN1677136A (zh) 2004-04-01 2005-10-05 佳能株式会社 衍射光学元件和包括所述衍射光学元件的光学系统
US7301702B2 (en) 2004-04-01 2007-11-27 Canon Kabushiki Kaisha Diffractive optical element and optical system including the same
US7612941B2 (en) 2004-04-01 2009-11-03 Canon Kabuhsiki Kaisha Diffractive optical element and optical system including the same
US7554735B2 (en) 2004-04-01 2009-06-30 Canon Kabushiki Kaisha Diffractive optical element and optical system including the same
US20050219702A1 (en) 2004-04-01 2005-10-06 Canon Kabushiki Kaisha Diffractive optical element and optical system including the same
US20080043335A1 (en) 2004-04-01 2008-02-21 Canon Kabushiki Kaisha Diffractive optical element and optical system including the same
US20070297057A1 (en) 2004-04-01 2007-12-27 Canon Kabushiki Kaisha Diffractive optical element and optical system including the same
US20060012772A1 (en) * 2004-07-15 2006-01-19 Mitutoyo Corporation Absolute distance measuring device
JP2006329897A (ja) 2005-05-30 2006-12-07 Tokyo Institute Of Technology 透明板に映る2重像を用いた距離計測方法
US20070086083A1 (en) * 2005-10-14 2007-04-19 Hon Hai Precision Industry Co., Ltd. Optical locating device
CN1844958A (zh) 2006-04-29 2006-10-11 杭州照相机械研究所 非球面衍射光学透镜
CN1963595A (zh) 2006-12-07 2007-05-16 四川大学 非平面光栅式三维自由立体显示屏
US20110286001A1 (en) * 2009-01-26 2011-11-24 Atsushi Taniguchi Dark-field defect inspecting method, dark-field defect inspecting apparatus, aberration analyzing method, and aberration analyzing apparatus
US20100322481A1 (en) 2009-06-19 2010-12-23 Canon Kabushiki Kaisha Image processing apparatus and method
JP2011002387A (ja) 2009-06-19 2011-01-06 Canon Inc 画像処理装置および画像処理方法
US20130286191A1 (en) * 2010-12-27 2013-10-31 Hitachi High-Technologies Corporation Inspection apparatus

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
International Search Report issued Dec. 6, 2011 in International (PCT) Application No. PCT/JP2011/006217.
Office Action mailed Sep. 2, 2014 along with a search report in corresponding Chinese patent application No. 201180005647.8 with English translation of search report.

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9572682B2 (en) 2011-09-29 2017-02-21 Arthromeda, Inc. System and method for precise prosthesis positioning in hip arthroplasty
US10314666B2 (en) 2011-09-29 2019-06-11 Arthromeda, Inc. System and method for precise prosthesis positioning in hip arthroplasty
US9597096B2 (en) 2013-03-15 2017-03-21 Arthromeda, Inc. Systems and methods for providing alignment in total knee arthroplasty
US9955983B2 (en) 2013-03-15 2018-05-01 Arthromeda, Inc. Systems and methods for providing alignment in total knee arthroplasty

Also Published As

Publication number Publication date
JPWO2012063463A1 (ja) 2014-05-12
US20120281086A1 (en) 2012-11-08
WO2012063463A1 (ja) 2012-05-18
CN102695939B (zh) 2015-06-17
CN102695939A (zh) 2012-09-26
JP5833546B2 (ja) 2015-12-16

Similar Documents

Publication Publication Date Title
US9121702B2 (en) Distance measurement apparatus and distance measurement method
US8711215B2 (en) Imaging device and imaging method
US9142582B2 (en) Imaging device and imaging system
JP5358039B1 (ja) 撮像装置
US9182602B2 (en) Image pickup device and rangefinder device
US8390799B2 (en) Method for measuring optical characteristics of diffraction optical element and apparatus for measuring optical characteristics of diffraction optical element
JP5144841B1 (ja) 撮像装置
KR200492211Y1 (ko) 근적외선 이미징 렌즈
US8284381B2 (en) Distance measuring apparatus and distance measuring method
US8836928B2 (en) Method for measuring wavefront aberration and wavefront aberration measuring apparatus
US9134126B2 (en) Image processing device, and image processing method
JP6729960B2 (ja) カメラモジュール調整装置及びカメラモジュール調整方法
JP2024028237A (ja) 分光カメラ、撮像方法、プログラム及び記録媒体
JP6376809B2 (ja) 三次元形状計測システムに用いられる投影装置および撮像装置
JP5264847B2 (ja) 測距装置、レンズシステムおよび撮像装置
JP2009151154A (ja) 受光素子、焦点検出装置および撮像装置
JP2015031754A (ja) 撮像装置及びその制御方法、情報処理装置及びその制御方法、及び、システム
US20200073117A1 (en) Chromatic lens and methods and systems using same
US20230090825A1 (en) Optical element assembly, optical apparatus, estimation method, and non-transitory storage medium storing estimation program
JP5546343B2 (ja) 測距装置及びそれを有する光学機器
JP6137839B2 (ja) 受信光学系
JP2017207720A (ja) 撮像光学系及び撮像装置
JP2005025140A (ja) 焦点検出装置

Legal Events

Date Code Title Description
AS Assignment

Owner name: PANASONIC CORPORATION, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:KIMURA, MASAYUKI;REEL/FRAME:028982/0499

Effective date: 20120531

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20230901