US20070222903A1 - Method and system for automated convergence and focus verification of projected images - Google Patents

Method and system for automated convergence and focus verification of projected images Download PDF

Info

Publication number
US20070222903A1
US20070222903A1 US11/803,796 US80379607A US2007222903A1 US 20070222903 A1 US20070222903 A1 US 20070222903A1 US 80379607 A US80379607 A US 80379607A US 2007222903 A1 US2007222903 A1 US 2007222903A1
Authority
US
United States
Prior art keywords
image
modulators
pulse
images
focus
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/803,796
Inventor
Randall Melton
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US11/803,796 priority Critical patent/US20070222903A1/en
Publication of US20070222903A1 publication Critical patent/US20070222903A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/12Picture reproducers
    • H04N9/31Projection devices for colour picture display, e.g. using electronic spatial light modulators [ESLM]
    • H04N9/3141Constructional details thereof
    • H04N9/317Convergence or focusing systems
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B7/00Mountings, adjusting means, or light-tight connections, for optical elements
    • G02B7/02Mountings, adjusting means, or light-tight connections, for optical elements for lenses
    • G02B7/04Mountings, adjusting means, or light-tight connections, for optical elements for lenses with mechanism for focusing or varying magnification
    • G02B7/08Mountings, adjusting means, or light-tight connections, for optical elements for lenses with mechanism for focusing or varying magnification adapted to co-operate with a remote control mechanism
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/12Picture reproducers
    • H04N9/16Picture reproducers using cathode ray tubes
    • H04N9/28Arrangements for convergence or focusing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/12Picture reproducers
    • H04N9/31Projection devices for colour picture display, e.g. using electronic spatial light modulators [ESLM]
    • H04N9/3191Testing thereof
    • H04N9/3194Testing thereof including sensor feedback
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/74Projection arrangements for image reproduction, e.g. using eidophor
    • H04N5/7416Projection arrangements for image reproduction, e.g. using eidophor involving the use of a spatial light modulator, e.g. a light valve, controlled by a video signal

Definitions

  • the present invention relates to the field of projection displays and more specifically to the automated measurement of such displays.
  • SLM spatial light modulator
  • FIG. 1 illustrates the convergence issue in a three-micromirror projection display (used for example only).
  • the three micromirrors each dedicated to one of the three primary colors of light (red, green, and blue), respectively are embedded within the optical system of the projector.
  • the images from these three micromirrors are combined by means of combining prisms and as a result, require mechanical alignment so that corresponding pixels from each array lay exactly on top of each other.
  • FIG. I a shows the same pixel from each of the red 1 , green 2 , and blue 3 micromirrors.
  • the red 1 and blue 3 pixels are shifted relative to the reference green 2 pixel as shown in Table 1 below.
  • FIG. 1 b shows the unaligned red 1 and blue 3 pixels relative to the reference green 2 pixel.
  • these show up as fuzzy edges in this B/W illustration, but as mis-aligned color pixels in a color photo.
  • the non-convergence is best observed along the edges of the pixel where a blue leading edge is seen at the top and right edge of the pixel and a red trailing edge is seen at the bottom and left edge of the pixel.
  • an operator would adjust the x and y locations of the red 1 and blue 3 micromirrors until the three images align with on another and the system is converged, resulting in a white image.
  • Focus is another parameter where the adjustment by an operator is often made subjectively. This parameter is more complicated to properly adjust, with many variables involved. For example, brightness can affect the focus significantly.
  • focus is usually accomplished by means of the projection lens, which can be either a zoom or fixed focal length lens.
  • FIG. 2 illustrates a row and column of pixels from a three-micromirror projection system, which is clearly out-of-focus.
  • the projector's operator will adjust the projection lens to provide the best focus, according to his desires.
  • FIG. 3 shows a row and column of pixels that have been both converged and focused manually by an operator. This shows the image properly converged, with the red, green, and blue pixels being properly aligned so as to appear as one pixel, white in color, and with sharp edges around both the pixel and around the hole in the center of the pixel. This hole in the center of the pixel is where the support post for the micromirror attaches to the mirror.
  • the method and system disclosed in this invention provide an objective tool for measuring the convergence and focus criteria of a projected image.
  • lens aberrations caused by lateral color shift are programmatically corrected.
  • snapshots are taken at several locations across the field-of-view. Data from each of these snapshots is separated into primary color images (typically red, green, and blue). The centroid of an identical row and column in each of the three (red, green, and blue) images is measured and the differences in the x and y position between the red (reference) centroid data and the green and blue data indicates the amount of adjustment of the green and blue images that is required to converge the image.
  • primary color images typically red, green, and blue
  • Focus for each primary color is accomplished by processing the three horizontal data arrays previously chosen by the user. After normalizing the data, a single-sided, scaled power spectrum of the array data is derived. Focus criteria are determined by summing the elements of the power spectrum array to the right of the first relative minima in the spectrum. This power spectrum sum is then maximized for optimal focus.
  • FIG. 1 a illustrates the three planes (red, green, and blue) for an out-of-convergence image. (prior art)
  • FIG. 1 b shows a row and column of non-converged pixels.
  • FIG. 2 shows a row and column of an out-of-focus image. (prior art)
  • FIG. 3 shows a row and column of a subjectively focused and converged image, based on operator's discretion.
  • FIGS. 4 a and 4 b are diagrams indicating where test images are taken in the image's field-of-view.
  • FIG. 4 c is a diagram of an un-converged image showing the x and y deltas (A).
  • FIG. 4 d is a diagram of a converged image.
  • FIG. 5 is a drawing showing how the horizontal and vertical waveforms for the selected row and column are generated.
  • FIG. 6 is a sketch of the differing red, green, and blue waveforms.
  • FIG. 7 describes the waveform's 90% amplitude level where the pulse width is measured.
  • FIG. 8 a indicates the desired waveform's centroid.
  • FIG. 8 b indicates a false waveform centroid.
  • FIG. 8 c illustrates the method for avoiding false waveform centroid measurements.
  • FIGS. 9 a, 9 b and 9 c illustrate the method of averaging the waveforms for multiple cuts across a pixel.
  • FIG. 10 a is the Fast Fourier Transform (FFT) for a horizontal pulse with sharp edges.
  • FIG. 10 b illustrates the power sum determined in the tail of the FFT.
  • FIG. 11 a shows an out-of-focus image.
  • FIG. 11 b shows an image focused using the method of this invention.
  • FIGS. 12 a and 12 b illustrate well-focused and poorly-focused waveforms, respectively.
  • FIGS. 12 c and 12 d show the waveforms of FIGS. 12 a and 12 b normalized to level 255.
  • FIG. 13 a is a block diagram of the automated convergence and focus system of this invention.
  • FIG. 13 b shows typical viewing window locations for the automated convergence and focus system of this invention.
  • FIG. 14 is a sketch of a typical monitor screen showing the automated convergence data of this invention.
  • FIG. 15 is a sketch of a typical monitor screen showing the automated focus data of this invention.
  • FIGS. 16 a and 16 b are diagrams of the data format used in the automated convergence and focus system of this invention.
  • FIG. 17 a is a portion of a flow chart showing the algorithm used for the automated focus and convergence operation.
  • FIG. 17 b is a portion of a flow chart showing the algorithm used for the automated focus and convergence operation.
  • FIG. 17 c is a portion of a flow chart showing the algorithm used for the automated focus and convergence operation.
  • FIG. 17 d is a portion of a flow chart showing the algorithm used for the automated focus and convergence operation.
  • FIG. 17 e is a portion of a flow chart showing the algorithm used for the automated focus and convergence operation.
  • the method and system of this invention provide an objective tool for measuring the convergence and focus criteria of a projected image.
  • lens aberrations caused by lateral color shift are programmatically corrected.
  • the method for objectively converging the primary color images involves capturing a magnified snapshot from several locations across the field-of-view of the picture and separating this data into a separate image for each of the modulators.
  • Each captured image typically is 640 ⁇ 480 pixel, 24-bit color image.
  • the captured images are separated into three 8-bit images, one for each modulator and typically are stored in DIB format.
  • the modulators typically each provide a primary color image, such as red, green, and blue images, simplifying the separation process.
  • this disclosure is in terms of the use of five 640 ⁇ 480 24-bit images, each dissolved into three 8-bit images, it should be understood this is for purposes of illustration and not for purposes of limitation. Other image resolutions, bit-depths, and numbers of images and modulators are also applicable to the processes taught herein.
  • a line and column of interest are chosen from the file and the resulting three horizontal (line) data arrays (Red, Green, and Blue) and three vertical (column) data arrays are used to determine the horizontal and vertical center-points of the three (Red, Green, and Blue) pixels.
  • the convergence adjustment is calculated by measuring the differences in the x and y dimensions between the Green and Blue pixel's center-points and the Red reference pixel's center-point. The green and blue center-points can then be moved to overlay the red center-point, thereby converge the image.
  • a row and column grid pattern is turned ON in the projected image, as shown in FIG. 4 a. For example, every 16th row and column of pixels might be turned ON.
  • a magnified 640 ⁇ 480 size image is then captured around one of the grid pattern intersections.
  • FIG. 4 b is a diagram showing the locations 41 - 45 where the five magnified 24-bit snapshots (A-E) are taken across the field-of-view 40 of the picture. These locations can vary, for example as shown by the dotted line squares 48 .
  • a line 46 and column 47 of data is chosen for each of the five snapshots for use in converging the picture. The conditions are established by turning ON only the pixels in the chosen row and column over the area of the snapshot.
  • FIG. 4 c is a diagram of an overlay of three un-converged red-green-blue images taken from the same row and column in the snapshot. Any of the three images could be used as a reference image, for example red in the diagram.
  • the x and y distances ( ⁇ Xg and ⁇ Yg) are then measured between the green row and column intersection 402 and the red (ref) intersection 401 as indicated and likewise the x and y distances ( ⁇ Xb and ⁇ Yb) are then measured between the blue row and column intersection 403 and the red (ref) intersection 401 .
  • the images from the green 402 and blue 403 SLM's can then be adjusted to overlay the image from the red 401 SLM as indicated in FIG. 4 d, where the red, green, and blue images are converged 405 .
  • FIG. 5 shows the details of the method used in converging the red, green, and blue images.
  • the pixel 59 of interest is chosen by the placement of the horizontal and vertical cursors (lines) 55 and 56 , respectively.
  • the data is then sampled to provide the horizontal row and vertical column waveforms 53 and 54 , respectively, for the selected pixel.
  • the 24-bit image is separated into three 8-bit images (red, green, and blue) that are processed individually.
  • the amplitude of the horizontal and vertical output signals 53 - 54 will go from 0 volts to a positive value (example: 0 to 5 volts) in the area of the selected pixel, as shown.
  • the method involves measuring the width of these two pulses (horizontal 53 and vertical 54 ) and then determining the center of each pulse.
  • the point where these two lines 55 - 56 intersect corresponds to the centroid of the pixel for a given color (red, green, or blue).
  • the red waveform 60 is the closest to being ideal and is therefore preferred as the reference to which the green and blue images are adjusted.
  • the green pulse 61 is slightly wider than the red pulse 60 .
  • the blue pulse 62 is also wider and tends to flare out even more at the lower levels. All these areas of complication have to be contended with in the process of converging the image.
  • FIG. 7 illustrates the method for overcoming the problems created by the amount of flaring of the three pulses at the lower levels.
  • the red 70 , green 71 , and blue 72 pulses are shown along with three lines 73 - 75 which represent the 10%, 50%, and 90% amplitude levels, respectively.
  • the method is to first normalize the three pulse heights so as to have the same amplitude (255 quantization level) and then to measure the pulse widths at the 90% amplitude level 75 . This places the point of measurement above most of the flaring and as a result provides accurate pulse widths. From these pulse widths the center of the three primary color pixels (red, green, and blue) is determined.
  • FIG. 8 a is a sketch of an ideal pulse 80 , which represents the pixel width, with the pulse width being measured at the 90% level 81 and the center of the pixel, indicated by line 82 , falling directly through the dip in the waveform.
  • the dip 803 at the top of the pulse 800 it is possible for the dip 803 at the top of the pulse 800 to fall to or below the 90% level 801 and as a result for the centerline 802 to be established in the center of one of the side lobes 804 instead of at the actual center of the pulse.
  • the method used to overcome this potential problem is described in FIG. 8 c.
  • the method for finding the center of the pulse is to first find the 90% level 811 on the leading edge, to go over the top of the pulse and down the trailing edge to the 10% level 812 , and then back up the trailing edge to the 90% level 813 .
  • the width of the pulse is then measured as the difference between the leading edge 90% level 811 and the trailing edge 90% level 813 .
  • the center of the pixel is shown by the line 814 at the mid-point of this difference. This approach avoids the possibility of making the measurements on one of the side lobes.
  • FIGS. 9 a & 9 b indicate how multiple scans are made across a pixel 900 with scans 902 being on one side of the pixel center hole 901 , other scans 903 through the area of the center hole 901 , and additional scans 904 on the other side of the pixel center hole 901 .
  • FIG. 9 c shows the results of averaging the scans.
  • waveforms 902 & 904 on either side of the pixel, do not exhibit a dip 905 at the peak amplitude while scans 903 through the center of the pixels do have the dip 905 at peak amplitude.
  • the average pulse 906 tends to reduce the effects of any flaring on the edges of the pulse across the pixel and reduces the size of the dip 907 at the top of the pulse 906 .
  • the convergence accuracy is improved by using this averaging approach.
  • Table 1 is an overview of the algorithm of this invention, used in converging the three SLM (red, green, and blue).
  • TABLE 1 CONVERGENCE ALGORITHM DETERMINE PIXEL WIDTH DETERMINE PIXEL HEIGHT DETERMINE PIXEL CENTER-X DETERMINE PIXEL CENTER-Y FIND HORIZONTAL LINE FIND VERTICAL COLUMN SET (ALIGN) LINE AND COLUMN
  • the method used for the automated focusing of a projected image, under varying illumination conditions, is very difficult. However, it is possible to adjust the focus of the optics to an optimal number during the assembly phase of a projector.
  • the method disclosed in this invention does this and can be used to assure that the focus parameter for shipped projector products are optimally focused and meet specification. The user of the projector can then manually focus the projector to match the brightness and other environmental conditions for a particular application.
  • focus for each color is accomplished by processing the three horizontal data arrays previously used in converging the pixels. After the data is normalized, a single-sided, scaled power spectrum of the data array is derived. Focus criteria are then determined by summing the elements of the power spectrum array to the right of the first relative minima in the spectrum. As the optics are adjusted, the value of the summed power spectrum is observed until a power sum maximum value is found.
  • FIG. 10 a illustrates a typical Fast Fourier Transform (FFT) 10 taken for the horizontal pulses with relative sharp edges, as shown earlier in FIG. 6 .
  • the power spectrum is the area 11 under the curve and is determined by adding the discrete values under the curve. By maximizing this sum the focus can be controlled. In practice, it was found that the sensitivity of the focus adjustment could be improved by maximizing the area 101 in the tail of the FFT curve 100 to the right of the first minima 102 , as shown in FIG. 10 b.
  • FIG. 11 a shows an out-of-focus image and FIG. 11 b shows the same image focused using the method of this invention.
  • FIGS. 12 a and 12 b show well focused and poorly focused, low light level pulses, respectively. Due to the sharp rise and fall times of the well focused pulse of FIG. 12 a, there is more area 101 in the tail 100 of the power spectrum ( FIG. 10 b ) curve then there is for the more rounded edges of the poorly pulses of FIG. 12 b. To reduce the effects of this problem, the pulses are first normalized to level 255 (maximum 8-bit level) before processing the data array, as shown in FIGS. 12 c and 12 d, respectively.
  • Table 2 is an overview of the algorithm of this invention, used in focusing the image. TABLE 2 FOCUS ALGORITHM USING CAPTURED DATA PERFORM SINGLE-SIDED FFT DETERMINE MAX-MIN FIND FIRST MINIMA OF ARRAY SUM ARRAY ELEMENTS TO RIGHT OF MINIMA
  • FIG. 13 a is a system block diagram for carrying out the convergence and focus methods of this invention.
  • Five Cameras 130 - 134 are used to store data from magnified views at the selected locations across the field-of-view; for example, locations at the upper left (UL) 1300 , upper right (UR) 1310 , lower left (LL) 1320 , lower right (LR) 1330 , and center (C) 1340 of the field, as indicated in FIG. 13 b.
  • the system is comprised of the cameras 130 - 134 , a video multiplexer (MUX) 135 , a frame grabber 136 , a computer 137 , and a viewing monitor 138 .
  • MUX video multiplexer
  • FIG. 14 illustrates a typical convergence screen 140 as seen by the operator on the monitor 138 .
  • This screen example shows the five sampling windows 141 - 145 . In each window, the center of the pixel height and width is displayed in windows 146 and 147 , respectively.
  • FIG. 15 illustrates a typical focus screen 150 as seen by the operator on the monitor 138 .
  • This screen example shows the pixel waveforms 151 - 155 for each pixel.
  • the power spectrum value is displayed for each pulse in a window (example, window 157 ).
  • Lights indicating the best focus 158 and Red (reference) focus 159 are also included.
  • FIG. 16 a shows the format for storing the data for each selected pixel in the computer's 137 memory.
  • the 24-bit (B, G, R) image is stored as a BMP file.
  • This file consists of a header 160 followed by the blue 1601 , green 1602 , and red 1603 data for horizontal pixel 0 through 639 ( 161 , 162 ) of line 0 ( 163 ). This process is repeated over and over for lines 1 through 479 ( 164 ).
  • the 24-bit data is then separated into the three R, G, B 8-bit data files, as shown in FIG. 16 b.
  • the file format of the data for each of the primary colors starts with a header 165 and a look-up-table (LUT) 166 .
  • the data then follows for pixel 0 through 639 for line 0 ( 167 ) through 479 ( 168 ).
  • the data from this system is used to converge and focus the red, green blue images. Aligning the three SLM's to provide proper convergence could be done using fly-in-place robots, or other automated techniques, or even by manual adjustment.
  • the optical focus is adjusted to provide a maximum power spectrum summation value.
  • FIGS. 17 a through 17 e provide a more detailed listing of the pseudo-code for the convergence and focus algorithm of this invention.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Optics & Photonics (AREA)
  • Video Image Reproduction Devices For Color Tv Systems (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)

Abstract

A method and system that objectively measures the convergence and focus of a 2 or 3 spatial light modulator (SLM) projection display. The system uses five (5) CCD cameras and a frame grabber to store red, green, and blue (R-G-B) data from selected pixels located in the corners and center of the projector's field-of-view. The horizontal and vertical locations for the R-G-B pixels at each of the five locations is determined and the delta (□) displacement of the green and blue pixels, relative to the reference red pixel, is calculated and used to converge the image. The optical focus of the system is also determined using a Fast Fourier Transform (FFT). The FFT is performed on this same data and a power spectrum summation beyond the first minima is determined. The focus is then adjusted to maximize this value.

Description

  • This application is a continuation of application Ser. No. 11/096,480, filed Apr. 01, 2005, which is a divisional of application Ser. No. 10/054,063, filed Nov. 13, 2001, which claims priority to Provisional Application No. 60/250,450, filed Nov. 30, 2000.
  • TECHNICAL FIELD
  • The present invention relates to the field of projection displays and more specifically to the automated measurement of such displays.
  • BACKGROUND
  • The convergence and focus of projection displays having more than one spatial light modulator (SLM) are typically determined subjectively by an operator. As a result, repeatability and tight tolerances in converging and focusing many projectors are difficult to accomplish. The results often depend on the skill and motivation of the person making the adjustments.
  • FIG. 1 illustrates the convergence issue in a three-micromirror projection display (used for example only). The three micromirrors, each dedicated to one of the three primary colors of light (red, green, and blue), respectively are embedded within the optical system of the projector. The images from these three micromirrors are combined by means of combining prisms and as a result, require mechanical alignment so that corresponding pixels from each array lay exactly on top of each other. FIG. I a shows the same pixel from each of the red 1, green 2, and blue 3 micromirrors. In this out-of-convergence example, where the green 2 pixel is the reference, the red 1 and blue 3 pixels are shifted relative to the reference green 2 pixel as shown in Table 1 below.
    TABLE 1
    X y
    R −0.4 −0.2
    G 0 0
    B +0.2 +0.2
  • It is clear from the figure that this system needs to be converged, at least in the area of the observed pixel. This is best illustrated by the picture of FIG. 1 b, which shows the unaligned red 1 and blue 3 pixels relative to the reference green 2 pixel. (Note: these show up as fuzzy edges in this B/W illustration, but as mis-aligned color pixels in a color photo). In the actual color picture, the non-convergence is best observed along the edges of the pixel where a blue leading edge is seen at the top and right edge of the pixel and a red trailing edge is seen at the bottom and left edge of the pixel. Typically, an operator would adjust the x and y locations of the red 1 and blue 3 micromirrors until the three images align with on another and the system is converged, resulting in a white image.
  • Focus is another parameter where the adjustment by an operator is often made subjectively. This parameter is more complicated to properly adjust, with many variables involved. For example, brightness can affect the focus significantly. In a projection system, focus is usually accomplished by means of the projection lens, which can be either a zoom or fixed focal length lens. FIG. 2 illustrates a row and column of pixels from a three-micromirror projection system, which is clearly out-of-focus. Typically, the projector's operator will adjust the projection lens to provide the best focus, according to his desires.
  • FIG. 3 shows a row and column of pixels that have been both converged and focused manually by an operator. This shows the image properly converged, with the red, green, and blue pixels being properly aligned so as to appear as one pixel, white in color, and with sharp edges around both the pixel and around the hole in the center of the pixel. This hole in the center of the pixel is where the support post for the micromirror attaches to the mirror.
  • What is needed is an objective method for convergence and focus criteria along with a measuring tool for implementing the method. This method needs to reflect the human element since the human eye is the final arbitrator in a display application. The invention disclosed herein addresses this need by means of both a method and a tool.
  • SUMMARY OF THE INVENTION
  • The method and system disclosed in this invention provide an objective tool for measuring the convergence and focus criteria of a projected image. In addition, lens aberrations caused by lateral color shift are programmatically corrected.
  • To converge the red, green, and blue images from a projector, snapshots are taken at several locations across the field-of-view. Data from each of these snapshots is separated into primary color images (typically red, green, and blue). The centroid of an identical row and column in each of the three (red, green, and blue) images is measured and the differences in the x and y position between the red (reference) centroid data and the green and blue data indicates the amount of adjustment of the green and blue images that is required to converge the image.
  • Focus for each primary color is accomplished by processing the three horizontal data arrays previously chosen by the user. After normalizing the data, a single-sided, scaled power spectrum of the array data is derived. Focus criteria are determined by summing the elements of the power spectrum array to the right of the first relative minima in the spectrum. This power spectrum sum is then maximized for optimal focus.
  • DESCRIPTION OF THE DRAWINGS
  • The included drawings are as follows:
  • FIG. 1 a illustrates the three planes (red, green, and blue) for an out-of-convergence image. (prior art)
  • FIG. 1 b shows a row and column of non-converged pixels. (prior art)
  • FIG. 2 shows a row and column of an out-of-focus image. (prior art)
  • FIG. 3 shows a row and column of a subjectively focused and converged image, based on operator's discretion. (prior art)
  • FIGS. 4 a and 4 b are diagrams indicating where test images are taken in the image's field-of-view.
  • FIG. 4 c is a diagram of an un-converged image showing the x and y deltas (A).
  • FIG. 4 d is a diagram of a converged image.
  • FIG. 5 is a drawing showing how the horizontal and vertical waveforms for the selected row and column are generated.
  • FIG. 6 is a sketch of the differing red, green, and blue waveforms.
  • FIG. 7 describes the waveform's 90% amplitude level where the pulse width is measured.
  • FIG. 8 a indicates the desired waveform's centroid.
  • FIG. 8 b indicates a false waveform centroid.
  • FIG. 8 c illustrates the method for avoiding false waveform centroid measurements.
  • FIGS. 9 a, 9 b and 9 c illustrate the method of averaging the waveforms for multiple cuts across a pixel.
  • FIG. 10 a is the Fast Fourier Transform (FFT) for a horizontal pulse with sharp edges.
  • FIG. 10 b illustrates the power sum determined in the tail of the FFT.
  • FIG. 11 a shows an out-of-focus image.
  • FIG. 11 b shows an image focused using the method of this invention.
  • FIGS. 12 a and 12 b illustrate well-focused and poorly-focused waveforms, respectively.
  • FIGS. 12 c and 12 d show the waveforms of FIGS. 12 a and 12 b normalized to level 255.
  • FIG. 13 a is a block diagram of the automated convergence and focus system of this invention.
  • FIG. 13 b shows typical viewing window locations for the automated convergence and focus system of this invention.
  • FIG. 14 is a sketch of a typical monitor screen showing the automated convergence data of this invention.
  • FIG. 15 is a sketch of a typical monitor screen showing the automated focus data of this invention.
  • FIGS. 16 a and 16 b are diagrams of the data format used in the automated convergence and focus system of this invention.
  • FIG. 17 a is a portion of a flow chart showing the algorithm used for the automated focus and convergence operation.
  • FIG. 17 b is a portion of a flow chart showing the algorithm used for the automated focus and convergence operation.
  • FIG. 17 c is a portion of a flow chart showing the algorithm used for the automated focus and convergence operation.
  • FIG. 17 d is a portion of a flow chart showing the algorithm used for the automated focus and convergence operation.
  • FIG. 17 e is a portion of a flow chart showing the algorithm used for the automated focus and convergence operation.
  • DETAILED DESCRIPTION OF ILLUSTRATIVE EMBODIMENTS
  • The method and system of this invention provide an objective tool for measuring the convergence and focus criteria of a projected image. In addition, lens aberrations caused by lateral color shift are programmatically corrected.
  • The method for objectively converging the primary color images, typically red, green, and blue, involves capturing a magnified snapshot from several locations across the field-of-view of the picture and separating this data into a separate image for each of the modulators.
  • While two or three captured images are enough to perform the convergence and focus operations, additional images improve the process and provide better results. Typically five captured images are used. Each captured image typically is 640×480 pixel, 24-bit color image. The captured images are separated into three 8-bit images, one for each modulator and typically are stored in DIB format. The modulators typically each provide a primary color image, such as red, green, and blue images, simplifying the separation process. Although this disclosure is in terms of the use of five 640×480 24-bit images, each dissolved into three 8-bit images, it should be understood this is for purposes of illustration and not for purposes of limitation. Other image resolutions, bit-depths, and numbers of images and modulators are also applicable to the processes taught herein.
  • After capturing the images, a line and column of interest are chosen from the file and the resulting three horizontal (line) data arrays (Red, Green, and Blue) and three vertical (column) data arrays are used to determine the horizontal and vertical center-points of the three (Red, Green, and Blue) pixels. Using the Red pixel (optional selection) as a reference, the convergence adjustment is calculated by measuring the differences in the x and y dimensions between the Green and Blue pixel's center-points and the Red reference pixel's center-point. The green and blue center-points can then be moved to overlay the red center-point, thereby converge the image.
  • In the method, a row and column grid pattern is turned ON in the projected image, as shown in FIG. 4 a. For example, every 16th row and column of pixels might be turned ON. A magnified 640×480 size image is then captured around one of the grid pattern intersections.
  • FIG. 4 b is a diagram showing the locations 41-45 where the five magnified 24-bit snapshots (A-E) are taken across the field-of-view 40 of the picture. These locations can vary, for example as shown by the dotted line squares 48. A line 46 and column 47 of data is chosen for each of the five snapshots for use in converging the picture. The conditions are established by turning ON only the pixels in the chosen row and column over the area of the snapshot.
  • For each of the five snapshots, a 24-bit DIB data file is separated into three 8-bit 640×480 data arrays, one representing each of the three primary colors, red, green, and blue. FIG. 4 c is a diagram of an overlay of three un-converged red-green-blue images taken from the same row and column in the snapshot. Any of the three images could be used as a reference image, for example red in the diagram. The x and y distances (ΔXg and ΔYg) are then measured between the green row and column intersection 402 and the red (ref) intersection 401 as indicated and likewise the x and y distances (ΔXb and ΔYb) are then measured between the blue row and column intersection 403 and the red (ref) intersection 401. The images from the green 402 and blue 403 SLM's can then be adjusted to overlay the image from the red 401 SLM as indicated in FIG. 4 d, where the red, green, and blue images are converged 405.
  • FIG. 5 shows the details of the method used in converging the red, green, and blue images. Given one of the snapshots 50, consisting of a magnified row 51 and column 52 of ON pixels, the pixel 59 of interest is chosen by the placement of the horizontal and vertical cursors (lines) 55 and 56, respectively. The data is then sampled to provide the horizontal row and vertical column waveforms 53 and 54, respectively, for the selected pixel. The 24-bit image is separated into three 8-bit images (red, green, and blue) that are processed individually. When the snapshot area is scanned for each of the three 8-bit images at the selected pixel location, the amplitude of the horizontal and vertical output signals 53-54 will go from 0 volts to a positive value (example: 0 to 5 volts) in the area of the selected pixel, as shown. The method involves measuring the width of these two pulses (horizontal 53 and vertical 54) and then determining the center of each pulse. The point where these two lines 55-56 intersect corresponds to the centroid of the pixel for a given color (red, green, or blue). In the case of a typical micromirror, there is a hole 57 located at the center of each pixel, where the mirror connects to its support post, which causes a dip 58 in the waveform. This dip 58 can complicate the process of locating the center of the waveform.
  • In locating the center of a row or column of pixels, there can be several complications involved. First, there is the dip at the center top of the waveform discussed above. Then there is the fact that waveforms representing the three colors each may have a somewhat different shape, as illustrated in FIG. 6. The red waveform 60 is the closest to being ideal and is therefore preferred as the reference to which the green and blue images are adjusted. The green pulse 61 is slightly wider than the red pulse 60. The blue pulse 62 is also wider and tends to flare out even more at the lower levels. All these areas of complication have to be contended with in the process of converging the image.
  • FIG. 7 illustrates the method for overcoming the problems created by the amount of flaring of the three pulses at the lower levels. The red 70, green 71, and blue 72 pulses are shown along with three lines 73-75 which represent the 10%, 50%, and 90% amplitude levels, respectively. The method is to first normalize the three pulse heights so as to have the same amplitude (255 quantization level) and then to measure the pulse widths at the 90% amplitude level 75. This places the point of measurement above most of the flaring and as a result provides accurate pulse widths. From these pulse widths the center of the three primary color pixels (red, green, and blue) is determined.
  • FIG. 8 a is a sketch of an ideal pulse 80, which represents the pixel width, with the pulse width being measured at the 90% level 81 and the center of the pixel, indicated by line 82, falling directly through the dip in the waveform. However, as shown in FIG. 8 b it is possible for the dip 803 at the top of the pulse 800 to fall to or below the 90% level 801 and as a result for the centerline 802 to be established in the center of one of the side lobes 804 instead of at the actual center of the pulse. The method used to overcome this potential problem is described in FIG. 8 c. Here three levels are determined for the pulse 810; i.e., (i) at the 90% level 811 on the leading edge, (ii) at the 10% level 812 on the trailing edge, and (iii) at the 90% level 813 on the trailing edge. The method for finding the center of the pulse is to first find the 90% level 811 on the leading edge, to go over the top of the pulse and down the trailing edge to the 10% level 812, and then back up the trailing edge to the 90% level 813. The width of the pulse is then measured as the difference between the leading edge 90% level 811 and the trailing edge 90% level 813. The center of the pixel is shown by the line 814 at the mid-point of this difference. This approach avoids the possibility of making the measurements on one of the side lobes.
  • To this point the discussion has centered around a single scan taken through the center of a pixel. In order to improve the accuracy, multiple sweeps (up to 20 passes) are taken across the pixel in both the horizontal and vertical direction and an average of these pulses is used to make the calculations, as described in FIG. 9.
  • FIGS. 9 a & 9 b indicate how multiple scans are made across a pixel 900 with scans 902 being on one side of the pixel center hole 901, other scans 903 through the area of the center hole 901, and additional scans 904 on the other side of the pixel center hole 901. These scans, shown for a row of pixels, apply equally to scans made across a column of pixels. For example, in the case where 20 scans are made per pixel, assume that
    a=1,
    a+m=10, and
    a+n+1=20.
  • FIG. 9 c shows the results of averaging the scans. Here, waveforms 902 & 904, on either side of the pixel, do not exhibit a dip 905 at the peak amplitude while scans 903 through the center of the pixels do have the dip 905 at peak amplitude. The average pulse 906 tends to reduce the effects of any flaring on the edges of the pulse across the pixel and reduces the size of the dip 907 at the top of the pulse 906. The convergence accuracy is improved by using this averaging approach.
  • Table 1 is an overview of the algorithm of this invention, used in converging the three SLM (red, green, and blue).
    TABLE 1
    CONVERGENCE ALGORITHM
    DETERMINE PIXEL WIDTH
    DETERMINE PIXEL HEIGHT
    DETERMINE PIXEL CENTER-X
    DETERMINE PIXEL CENTER-Y
    FIND HORIZONTAL LINE
    FIND VERTICAL COLUMN
    SET (ALIGN) LINE AND COLUMN
  • The method used for the automated focusing of a projected image, under varying illumination conditions, is very difficult. However, it is possible to adjust the focus of the optics to an optimal number during the assembly phase of a projector. The method disclosed in this invention does this and can be used to assure that the focus parameter for shipped projector products are optimally focused and meet specification. The user of the projector can then manually focus the projector to match the brightness and other environmental conditions for a particular application.
  • In the automated focus method disclosed herein, focus for each color (red, green, and blue) is accomplished by processing the three horizontal data arrays previously used in converging the pixels. After the data is normalized, a single-sided, scaled power spectrum of the data array is derived. Focus criteria are then determined by summing the elements of the power spectrum array to the right of the first relative minima in the spectrum. As the optics are adjusted, the value of the summed power spectrum is observed until a power sum maximum value is found.
  • FIG. 10 a illustrates a typical Fast Fourier Transform (FFT) 10 taken for the horizontal pulses with relative sharp edges, as shown earlier in FIG. 6. The power spectrum is the area 11 under the curve and is determined by adding the discrete values under the curve. By maximizing this sum the focus can be controlled. In practice, it was found that the sensitivity of the focus adjustment could be improved by maximizing the area 101 in the tail of the FFT curve 100 to the right of the first minima 102, as shown in FIG. 10 b. FIG. 11 a shows an out-of-focus image and FIG. 11 b shows the same image focused using the method of this invention.
  • As illustrated by FIG. 12, focus is dependent on the light levels. FIGS. 12 a and 12 b show well focused and poorly focused, low light level pulses, respectively. Due to the sharp rise and fall times of the well focused pulse of FIG. 12 a, there is more area 101 in the tail 100 of the power spectrum (FIG. 10 b) curve then there is for the more rounded edges of the poorly pulses of FIG. 12 b. To reduce the effects of this problem, the pulses are first normalized to level 255 (maximum 8-bit level) before processing the data array, as shown in FIGS. 12 c and 12 d, respectively.
  • Table 2 is an overview of the algorithm of this invention, used in focusing the image.
    TABLE 2
    FOCUS ALGORITHM
    USING CAPTURED DATA
    PERFORM SINGLE-SIDED FFT
    DETERMINE MAX-MIN
    FIND FIRST MINIMA OF ARRAY
    SUM ARRAY ELEMENTS TO RIGHT OF MINIMA
  • FIG. 13 a is a system block diagram for carrying out the convergence and focus methods of this invention. Five Cameras 130-134 are used to store data from magnified views at the selected locations across the field-of-view; for example, locations at the upper left (UL) 1300, upper right (UR) 1310, lower left (LL) 1320, lower right (LR) 1330, and center (C) 1340 of the field, as indicated in FIG. 13 b. The system is comprised of the cameras 130-134, a video multiplexer (MUX) 135, a frame grabber 136, a computer 137, and a viewing monitor 138.
  • FIG. 14 illustrates a typical convergence screen 140 as seen by the operator on the monitor 138. This screen example shows the five sampling windows 141-145. In each window, the center of the pixel height and width is displayed in windows 146 and 147, respectively.
  • FIG. 15 illustrates a typical focus screen 150 as seen by the operator on the monitor 138. This screen example shows the pixel waveforms 151-155 for each pixel. The power spectrum value is displayed for each pulse in a window (example, window 157). Lights indicating the best focus 158 and Red (reference) focus 159 are also included.
  • FIG. 16 a shows the format for storing the data for each selected pixel in the computer's 137 memory. First, the 24-bit (B, G, R) image is stored as a BMP file. This file consists of a header 160 followed by the blue 1601, green 1602, and red 1603 data for horizontal pixel 0 through 639 (161, 162) of line 0 (163). This process is repeated over and over for lines 1 through 479 (164). The 24-bit data is then separated into the three R, G, B 8-bit data files, as shown in FIG. 16 b. The file format of the data for each of the primary colors starts with a header 165 and a look-up-table (LUT) 166. The data then follows for pixel 0 through 639 for line 0 (167) through 479 (168).
  • In operation, the data from this system is used to converge and focus the red, green blue images. Aligning the three SLM's to provide proper convergence could be done using fly-in-place robots, or other automated techniques, or even by manual adjustment. The optical focus is adjusted to provide a maximum power spectrum summation value.
  • FIGS. 17 a through 17 e provide a more detailed listing of the pseudo-code for the convergence and focus algorithm of this invention.
  • The same techniques described herein for a 3-SLM application apply as well to a 2-SLM system.
  • While this invention has been described in the context of preferred embodiments, it will be apparent to those skilled in the art that the present invention may be modified in numerous ways and may assume embodiments other than that specifically set out and described above. Accordingly, it is intended by the appended claims to cover all modifications of the invention that fall within the true spirit and scope of the invention.

Claims (20)

1. A method for convergence of an image projected by at least two modulators, the method comprising:
turning on a test pixel in the projected image;
capturing a combined image of the test pixel;
separating the combined image into a separate image for each of the modulators;
determining a centroid of x and y pulses representing the test pixel's width and height for each of the separate images;
calculating a convergence misalignment of the separate images based on a difference in x and y position between the centroids; and
mechanically aligning the modulators based on the convergence misalignment.
2. The method of claim 1 wherein the projected image is generated by at least three spatial light modulators.
3. The method of claim 2 wherein the spatial light modulators are micromirror devices.
4. The method of claim 3 wherein there are three micromirror devices projecting a red image, a green image, and a blue image as the separate images.
5. The method of claim 4, wherein the mechanically aligning the modulators further comprises adjusting the micromirror devices projecting the green and blue images to align with the micromirror device projecting the red image.
6. The method of claim 1 wherein the capturing the combined image further comprises:
capturing a first image near a center of the projected image; and
capturing at least four additional images near a perimeter of the projected image.
7. The method of claim 1 wherein the capturing the combined image further comprises:
taking multiple scans in both horizontal and vertical directions across the test pixel; and
averaging the multiple scans to generate the combined image.
8. The method of claim 7, wherein the taking the multiple scans comprises taking up to 20 scans.
9. The method of claim 1, wherein the determining the centroids further comprises normalizing the height of the x and y pulses for each of the separate images.
10. The method of claim 9 wherein the determining the centroids further comprises measuring the test pixel width at a 90% level of the normalized height of the x and y pulses for each of the separated images.
11. The method of claim 10, wherein the determining and measuring further comprise, for each of the centroids:
locating 90% levels for the x pulse's leading and trailing edges;
locating 90% levels for the y pulse's leading and trailing edges; and
setting the centroid as an intersection of a mid-point between the 90% levels of the x pulse's leading and trailing edges and a mid-point between the 90% levels of the y pulse's leading and trailing edges.
12. The method of claim 9, wherein the determining the centroids further comprises, for both the x and y pulses for each of the separate images:
locating a first location that is greater than a first threshold on a leading edge of the normalized pulse;
locating a second location that is lower than a second threshold on a trailing edge of the normalized pulse;
locating a third location that exceeds a third threshold on the trailing edge of the normalized pulse, wherein the third location is between the first and second locations; and
averaging the first and third locations.
13. The method of claim 12, wherein the locating the third location is performed after the locating the second location.
14. The method of claim 12, wherein the first and third thresholds are 90% of the normalized height of the pulse, and the second threshold is 10% of the normalized height of the pulse.
15. The method of claim 1, wherein the combined image is a 24-bit image, and wherein each of the separate images is an 8-bit image.
16. The method of claim 1, wherein the mechanically aligning the modulators is performed automatically using fly-in-place robotics.
17. The method of claim 1, wherein the mechanically aligning the modulators is performed manually.
18. The method of claim 1, wherein the captured combined image and the separated images are magnified from the projected image.
19. An apparatus comprising:
an image capture device for capturing a combined image of a test pixel projected by at least two modulators; and
a processor coupled to the image capture device for receiving the captured combined image, the processor configured to
separate the combined image into a separate image for each of the modulators,
determine a centroid of x and y pulses representing the test pixel's width and height for each of the separate images,
calculate a convergence misalignment of the separate images based on a difference in x and y position between the centroids, and
provide data to mechanically align the modulators based on the convergence misalignment.
20. The apparatus of claim 19, wherein the processor is configured to automatically align the modulators.
US11/803,796 2000-11-30 2007-05-16 Method and system for automated convergence and focus verification of projected images Abandoned US20070222903A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US11/803,796 US20070222903A1 (en) 2000-11-30 2007-05-16 Method and system for automated convergence and focus verification of projected images

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US25045000P 2000-11-30 2000-11-30
US10/054,063 US6995810B2 (en) 2000-11-30 2001-11-13 Method and system for automated convergence and focus verification of projected images
US11/096,480 US7268837B2 (en) 2000-11-30 2005-04-01 Method and system for automated convergence and focus verification of projected images
US11/803,796 US20070222903A1 (en) 2000-11-30 2007-05-16 Method and system for automated convergence and focus verification of projected images

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US11/096,480 Continuation US7268837B2 (en) 2000-11-30 2005-04-01 Method and system for automated convergence and focus verification of projected images

Publications (1)

Publication Number Publication Date
US20070222903A1 true US20070222903A1 (en) 2007-09-27

Family

ID=26732585

Family Applications (3)

Application Number Title Priority Date Filing Date
US10/054,063 Expired - Lifetime US6995810B2 (en) 2000-11-30 2001-11-13 Method and system for automated convergence and focus verification of projected images
US11/096,480 Expired - Lifetime US7268837B2 (en) 2000-11-30 2005-04-01 Method and system for automated convergence and focus verification of projected images
US11/803,796 Abandoned US20070222903A1 (en) 2000-11-30 2007-05-16 Method and system for automated convergence and focus verification of projected images

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US10/054,063 Expired - Lifetime US6995810B2 (en) 2000-11-30 2001-11-13 Method and system for automated convergence and focus verification of projected images
US11/096,480 Expired - Lifetime US7268837B2 (en) 2000-11-30 2005-04-01 Method and system for automated convergence and focus verification of projected images

Country Status (1)

Country Link
US (3) US6995810B2 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140285780A1 (en) * 2008-07-30 2014-09-25 Seiko Epson Corporation Projector and control method of projector

Families Citing this family (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7324279B2 (en) * 2000-12-28 2008-01-29 Texas Instruments Incorporated Dual modulator projection system
US7227686B1 (en) 2002-01-22 2007-06-05 General Photonics Corporation Tunable PMD emulators and compensators
JP3742027B2 (en) * 2002-04-08 2006-02-01 Necビューテクノロジー株式会社 Projection image distortion correction method, distortion correction program, and projection-type image display device
US7154659B1 (en) * 2002-04-18 2006-12-26 General Photonics Corporation Optical depolarizers and DGD generators based on optical delay
US7215362B2 (en) * 2002-10-31 2007-05-08 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Auto-calibration of multi-projector systems
US7391977B2 (en) 2003-03-12 2008-06-24 General Photonics Corporation Monitoring mechanisms for optical systems
US7796894B1 (en) 2003-07-30 2010-09-14 General Photonics Corporation Reduction of noise and polarization mode dispersion (PMD) based on optical polarization stabilizer in fiber transmission
US7522785B2 (en) * 2004-12-01 2009-04-21 General Photonics Corporation Measurements of polarization-dependent loss (PDL) and degree of polarization (DOP) using optical polarization controllers
US20070030452A1 (en) * 2005-08-08 2007-02-08 N-Lighten Technologies Image adaptation system and method
US7693419B1 (en) 2005-11-23 2010-04-06 General Photonics Corporation Optical spectrum analysis using optical interferometry
US8073326B2 (en) * 2006-12-06 2011-12-06 General Photonics Corporation Optical polarization division multiplexing in optical communication
US7952711B1 (en) 2007-03-26 2011-05-31 General Photonics Corporation Waveplate analyzer based on multiple tunable optical polarization rotators
US7723670B1 (en) 2007-03-26 2010-05-25 General Photonics Corporation Optical differential group delay module with folded optical path
US7945130B2 (en) * 2007-11-15 2011-05-17 General Photonics Corporation Mode scrambling apparatus for multimode fiber
US8422882B1 (en) 2008-02-04 2013-04-16 General Photonics Corporation Monitoring polarization-mode dispersion and signal-to-noise ratio in optical signals based on polarization analysis
US8345238B2 (en) * 2008-02-04 2013-01-01 General Photonics Corporation Measuring optical spectral property of light based on polarization analysis
TWI353531B (en) * 2008-05-13 2011-12-01 Wistron Corp Method and a device of adjusting a color temperatu
US7901095B2 (en) * 2009-03-27 2011-03-08 Seiko Epson Corporation Resolution scalable view projection
EP3657228A1 (en) * 2009-07-09 2020-05-27 Howard Hughes Medical Institute Microscopy with adaptive optics
US8953049B2 (en) * 2010-11-24 2015-02-10 Echostar Ukraine L.L.C. Television receiver—projector compensating optical properties of projection surface
US8866107B2 (en) 2011-01-19 2014-10-21 Howard Hughes Medical Institute Wavefront compensation for deep tissue optical microscopy
WO2013010151A1 (en) 2011-07-14 2013-01-17 Howard Hughes Medical Institute Microscopy with adaptive optics
US8780433B2 (en) 2011-09-28 2014-07-15 General Photonics Corporation Polarization scrambling based on cascaded optical polarization devices having modulated optical retardation
KR101807159B1 (en) * 2015-12-04 2017-12-11 카페24 주식회사 Method, Apparatus and System for Transmitting Video Based On Multiple Cameras
US10276075B1 (en) * 2018-03-27 2019-04-30 Christie Digital System USA, Inc. Device, system and method for automatic calibration of image devices
CN108683898B (en) * 2018-05-31 2020-10-09 歌尔股份有限公司 Method, device and equipment for correcting color convergence errors
US10694160B2 (en) 2018-11-20 2020-06-23 Disney Enterprises, Inc. Channel based projector calibration

Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5258830A (en) * 1990-03-16 1993-11-02 Electrohome Limited Automatic video convergence system
US5345262A (en) * 1992-07-31 1994-09-06 Hughes-Jvc Technology Corporation Automatic convergence system for color video projector
US5532765A (en) * 1993-03-17 1996-07-02 Matsushita Electric Industrial Co., Ltd. Image correction apparatus using a displayed test signal
US5699440A (en) * 1993-12-02 1997-12-16 Genop Ltd. Method and system for testing the performance of at least one electro-optical test device
US5835135A (en) * 1995-03-08 1998-11-10 Minolta Co., Ltd. Device for measuring a glow center of a display device
US20020024708A1 (en) * 1999-08-05 2002-02-28 Microvision, Inc. Scanned imaging apparatus with switched feeds
US6424412B1 (en) * 2000-08-30 2002-07-23 Sony Corporation Efficient system and method for detecting and correcting laser misalignment of plural laser beams
US6456339B1 (en) * 1998-07-31 2002-09-24 Massachusetts Institute Of Technology Super-resolution display
US6483555B1 (en) * 1996-06-12 2002-11-19 Barco N.V. Universal device and use thereof for the automatic adjustment of a projector
US6503195B1 (en) * 1999-05-24 2003-01-07 University Of North Carolina At Chapel Hill Methods and systems for real-time structured light depth extraction and endoscope using real-time structured light depth extraction
US6717625B1 (en) * 1997-12-01 2004-04-06 Barco N.V. Method and device for adjusting one or more projectors
US20050162572A1 (en) * 1999-05-26 2005-07-28 Olympus Optical Co., Ltd. Color reproduction system for making color display of four or more primary colors based on input tristimulus values
US6965396B1 (en) * 1995-06-28 2005-11-15 The Johns Hopkins University Video-centroid integrated circuit
US7009651B2 (en) * 2000-10-12 2006-03-07 Amnis Corporation System and method for high numeric aperture imaging systems
US7457458B1 (en) * 1999-11-26 2008-11-25 Inb Vision Ag. Method and apparatus for defining and correcting image data

Patent Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5258830A (en) * 1990-03-16 1993-11-02 Electrohome Limited Automatic video convergence system
US5345262A (en) * 1992-07-31 1994-09-06 Hughes-Jvc Technology Corporation Automatic convergence system for color video projector
US5532765A (en) * 1993-03-17 1996-07-02 Matsushita Electric Industrial Co., Ltd. Image correction apparatus using a displayed test signal
US5699440A (en) * 1993-12-02 1997-12-16 Genop Ltd. Method and system for testing the performance of at least one electro-optical test device
US5835135A (en) * 1995-03-08 1998-11-10 Minolta Co., Ltd. Device for measuring a glow center of a display device
US6965396B1 (en) * 1995-06-28 2005-11-15 The Johns Hopkins University Video-centroid integrated circuit
US6483555B1 (en) * 1996-06-12 2002-11-19 Barco N.V. Universal device and use thereof for the automatic adjustment of a projector
US6717625B1 (en) * 1997-12-01 2004-04-06 Barco N.V. Method and device for adjusting one or more projectors
US6456339B1 (en) * 1998-07-31 2002-09-24 Massachusetts Institute Of Technology Super-resolution display
US6503195B1 (en) * 1999-05-24 2003-01-07 University Of North Carolina At Chapel Hill Methods and systems for real-time structured light depth extraction and endoscope using real-time structured light depth extraction
US20050162572A1 (en) * 1999-05-26 2005-07-28 Olympus Optical Co., Ltd. Color reproduction system for making color display of four or more primary colors based on input tristimulus values
US20020024708A1 (en) * 1999-08-05 2002-02-28 Microvision, Inc. Scanned imaging apparatus with switched feeds
US7457458B1 (en) * 1999-11-26 2008-11-25 Inb Vision Ag. Method and apparatus for defining and correcting image data
US6424412B1 (en) * 2000-08-30 2002-07-23 Sony Corporation Efficient system and method for detecting and correcting laser misalignment of plural laser beams
US7009651B2 (en) * 2000-10-12 2006-03-07 Amnis Corporation System and method for high numeric aperture imaging systems

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140285780A1 (en) * 2008-07-30 2014-09-25 Seiko Epson Corporation Projector and control method of projector

Also Published As

Publication number Publication date
US20020093631A1 (en) 2002-07-18
US7268837B2 (en) 2007-09-11
US6995810B2 (en) 2006-02-07
US20050168659A1 (en) 2005-08-04

Similar Documents

Publication Publication Date Title
US7268837B2 (en) Method and system for automated convergence and focus verification of projected images
EP0498659B1 (en) Adjustment apparatus for video display means
US5200815A (en) Convergence correction apparatus for a video projection system
EP1861748B1 (en) Method of and apparatus for automatically adjusting alignement of a projector with respect to a projection screen
US6950139B2 (en) Image reading device and storage medium storing control procedure for image reading device
US6198514B1 (en) Color misconvergence measurement using a common monochrome image
EP3544296A1 (en) Projection control apparatus, projection apparatus, projection control method, and program
JP2606185B2 (en) Method for measuring registration of solid-state image sensor
JP2000241874A (en) Method and device for automatically adjusting screen position for projector
JPH0614750B2 (en) Color original imaging device
US5796425A (en) Elimination of the effect of difference in vertical scanning frequency between a display and a camera imaging the display
US6922483B2 (en) Methods for measuring DMD low frequency spatial uniformity
US7196741B2 (en) Automatic image convergence for projection television systems
CN111179815B (en) Method for collecting and correcting normal brightness and chromaticity of LED display module
CN108848358B (en) Method and device for correcting color convergence errors
JPH0969973A (en) Position adjusting method for solid-state image pickup element
KR100471084B1 (en) Image processing system and method
JPH07131802A (en) Image correcting device
JPH05308554A (en) Displacement detector for solid-state image pickup device
JP3853674B2 (en) Convergence adjustment system and convergence adjustment method
JP3859245B2 (en) How to center the chart
JPH11196316A (en) Image reader
JP2005124105A (en) Projector and color correcting method thereof
JP2000172846A (en) Image processor
JP2000324517A (en) Adhering chart for linear sensor camera element and its adhering method

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION