WO2000011873A1 - Method and apparatus for inspection of printed circuit boards - Google Patents
Method and apparatus for inspection of printed circuit boards Download PDFInfo
- Publication number
- WO2000011873A1 WO2000011873A1 PCT/IL1999/000450 IL9900450W WO0011873A1 WO 2000011873 A1 WO2000011873 A1 WO 2000011873A1 IL 9900450 W IL9900450 W IL 9900450W WO 0011873 A1 WO0011873 A1 WO 0011873A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- sensors
- image
- subsystem
- article
- sensor
- Prior art date
Links
- 0 C*1(CCC1)N Chemical compound C*1(CCC1)N 0.000 description 2
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N21/00—Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
- G01N21/84—Systems specially adapted for particular applications
- G01N21/88—Investigating the presence of flaws or contamination
- G01N21/95—Investigating the presence of flaws or contamination characterised by the material or shape of the object to be examined
- G01N21/956—Inspecting patterns on the surface of objects
Definitions
- the present invention relates to article inspection systems and methods generally and more particularly to systems and methods for inspecting generally two dimensional articles, such as printed circuit boards
- an image acquisition system including a plurality of sensors each including a multiplicity of sensor elements, a pre-scan calibration subsystem, employing a predetermined test pattern, which is sensed by the plurality of sensors, for providing an output indication in two dimensions of distortions in the output of the plurality of sensors, the output indication being employed to generate a function which maps the locations viewed by the sensor elements in at least two dimensions, and a distortion correction subsystem operative during scanning of an article by the plurality of sensors to correct the distortions by employing the output indication
- the plurality of sensors include plural sensors having different spectral sensitivities
- the distortion co ⁇ ection subsystem performs non-zero'th order interpolation of pixels in the outputs of the plurality of sensors
- the plurality of sensors include sensors having differing spectral sensitivities and the function is dependent on differing accumulation times employed for the sensors having differing spectral sensitivities
- the distortion correction subsystem compensates for variations in pixel shape in the plurality of sensors
- the distortion correction subsystem is operative to an accuracy of better than 5% of pixel size of the multiplicity of sensor elements
- an image acquisition system including a plurality of sensors each including a multiplicity of sensor elements, a pre-scan calibration subsystem, employing a predetermined test pattern, which is sensed by the plurality of sensors, for providing an output indication of distortions in the output of the plurality of sensors, the output indication being employed to generate a correction function, and a distortion correction subsystem operative during scanning of an article by the plurality of sensors to correct the distortions by employing the correction function, the distortion correction subsystem being operative to an accuracy of better than 5% of pixel size of the multiplicity of sensor elements
- an image acquisition system including a plurality of sensors, a pre-scan calibration subsystem, employing a predetermined test pattern, which is sensed by the plurality of sensors while being moved relative to the plurality of sensors in a direction of relative movement, for providing an output indication of distortions in the output of the plurality of sensors, the pre-scan calibration system being operative to correlate images of at least one target on the test pattern as seen by the plurality of sensors, thereby to determine the relative orientation of the plurality of sensors, and a distortion correction subsystem operative to correct the distortions by employing the output indication
- the pre-scan calibration subsystem also is operative to provide an output indication of the o ⁇ entation of the plurality of sensors relative to the scan direction
- each of the plurality of sensors includes a multiplicity of sensor elements
- the pre-scan calibration subsystem also is operative to determine the pixel size characteristic of each of the multiplicity of sensor elements of each of the plurality of sensors
- the pre-scan calibration subsystem is operative to determine the pixel size characteristic of each of the multiplicity of sensor elements of each of the plurality of sensors by causing the plurality of sensors to view a grid formed of a multiplicity of parallel uniformly spaced lines, formed on the test pattern
- an image acquisition system including a plurality of sensors each including a multiplicity of sensor elements, a pre-scan calibration subsystem, employing a predetermined test pattern, which is sensed by the plurality of sensors, for providing an output indication in two dimensions of distortions in the output of the plurality of sensors, the output indication being employed to generate a function which maps the locations viewed by the sensor elements in at least two dimensions, and an distortion correction subsystem operative during scanning of an article by the plurality of sensors to correct the distortions by employing the output indication
- an article inspection system including an image acquisition subsystem operative to acquire an image of an article to be inspected, an image analysis subsystem for identifying at least one predetermined characteristic of the article from the image, and an output indication subsystem for providing an output indication of the presence of the at last one predetermined characteristic of the article,
- the camera assembly includes a plurality of sensor assemblies, self calibration apparatus for determining a geometrical relationship between the sensor assemblies, and sensor output modification apparatus for modifying outputs of the plurality of sensor assemblies based on the geometrical relationship between the sensor assemblies, the sensor output modification apparatus including electronic interpolation apparatus operative to perform non-zero'th order interpolation of pixels in the outputs of the plurality of sensor assemblies
- an article inspection system including an image acquisition subsystem operative to acquire an image of an article to be inspected, an image analysis subsystem for identifying at least one predetermined characteristic of the article from the image, and an output indication subsystem for providing an output indication of the presence of the at least one predetermined characteristic of the article,
- the camera assembly includes a plurality of sensor assemblies, self calibration apparatus for determining a geometrical relationship between the sensor assemblies, and sensor output modification apparatus for modifying outputs of the plurality of sensor assemblies based on the geometrical relationship between the sensor assemblies, the sensor output modification apparatus being operative to modify the outputs of the plurality of sensor assemblies to sub- pixel accuracy
- an article inspection system including an image acquisition subsystem operative to acquire an image of an article to be inspected, an image analysis subsystem for identifying at least one predetermined characteristic of the article from the image, and an output indication subsystem for providing an output indication of the presence of the at least one predetermined characteristic of the article, characterized in the camera assembly includes at least one sensor assembly, and sensor output modification apparatus for modifying at least one output of the at least one sensory assembly based at least in part on an optical distortion associated with the at least one sensor assembly
- the optical distortion includes magnification distortion
- the optical distortion includes chromatic aberration
- an article inspection system including an image acquisition subsystem operative to acquire an image of an article to be inspected, an image analysis subsystem for identifying at least one predetermined characteristic of the article from the image, and an output indication subsystem for providing an output indication of the presence of the at least one predetermined characteristic of the article, characterized in the camera assembly includes at least one sensor assembly, sensor output modification apparatus for modifying at least one output of the at least one sensory assembly, the sensor output modification apparatus including a function generator which generates a function which maps locations on the sensor assembly to a collection of scan locations
- an article inspection system including a camera assembly operative to acquire an image of an article to be inspected, an image analysis subsystem for identifying at least one predetermined characteristic of the article from the image, and an output indication subsystem for providing an output indication of the presence of the at last one predetermined characteristic of the article, characterized in the camera assembly includes a user interface which enables a user to select resolution of the image acquired by the camera assembly, an electro- optical sensor assembly, and an electronic resolution modifier operative downstream of the electro-optical sensor assembly
- Fig 1 is a simplified block diagram illustration of an article inspection system constructed and operative in accordance with a preferred embodiment of the present invention
- Fig 2 is a simplified illustration of parts of a preferred test pattern, certain portions of which are not drawn to scale, along with a simplified indication of the fields of view of individu ⁇ l line sensors viewing the test pattern,
- Fig 3 is a simplified block diagram illustration of mapping function generator circuitry forming part of the system of Fig 1
- Fig 4A is a simplified flow chart illustrating operation of pixel size and shape determination functionality of the mapping function generator circuitry of
- Fig 4B is a simplified illustration of geometrical distortion addressed by the functionality of Fig 4A
- Fig 4C is a simplified semi-pictorial, semi-graphical illustration of the functionality described in the flowchart of Fig 4A
- Fig 5A is a simplified flow chart illustrating operation of test pattern angle determination functionality of the mapping function generator circuitry of Fig
- Fig 5B is a simplified illustration of the geometrical distortion addressed by the functionality of Fig 5 A
- Fig 5C is a simplified semi-pictorial, semi-graphical illustration of the functionality described in the flowchart of Fig 5 A
- Fig 6A is a simplified flow chart illustrating operation of Determination of Relative Orientation of Sensors by Correlating Images of Test Pattern Targets at Edges of the Field of View of Sensors functionality of the mapping function generator circuitry of Fig 3,
- Fig 6B is a simplified illustration of the geometrical distortion addressed by the functionality of Fig 6 A,
- Fig 6C is a simplified semi-picto ⁇ al, semi-graphical illustration of the functionality described in the flowchart of Fig 6A,
- Fig 7A is a simplified flow chart illustrating operation of Determination of X Overlap and Y Offset of Sensors by Correlation of Adjacent Images functionality of the mapping function generator circuitry of Fig 3,
- Fig 7B is a simplified illustration of the geometrical distortion addressed by the functionality of Fig 7 A
- Fig 7C is a simplified illustration of the functionality described in the flowchart of Fig 7 A.
- Fig 8A is a simplified flow chart illustrating operation of Determination of X and Y Offsets for Multiple Colors functionality of the mapping function generator circuitry of Fig 3,
- Fig 8B is a simplified illustration of the geometrical distortion addressed by the functionality of Fig 8 A,
- Fig 8C is a simplified illustration of the functionality described in the flowchart of Fig 8 A, Figs 9A and 9B, taken together, are simplified flowchart illustrations of a preferred method of implementing mapping function generator circuitry forming part of the system of Fig 1 ,
- Fig 10A is a simplified illustration of acquisition of a target image by multiple cameras under ideal conditions
- Fig 10B is a simplified illustration of image buffers for storing the acquired image of 10 A
- Fig 1 1A is a simplified illustration of multiple cameras acquiring an image of a target where the camera fields of view are mutually skewed and overlapped
- Fig 1 IB is a simplified illustration of image buffers for storing the acquired image of 1 1 A
- Fig 12 is a simplified illustration useful in understanding Y- resamphng functionality of the image correction circuitry 110 of Fig 1 ,
- Figs 13 and 14, taken together, are simplified illustrations useful in understanding overlap correction functionality of the image correction circuitry 1 10
- Figs 15 and 16 taken together, are simplified illustration useful in understanding X-resamphng functionality of the image correction circuitry 1 10 of
- Fig 17 is a simplified illustration useful in understanding aspects of accumulation shift correction functionality of the image correction circuitry 1 10 of
- Fig 1 is a simplified block diagram illustration of an inspection system constructed and operative in accordance with a preferred embodiment of the present invention
- the inspection system of Fig 1 comprises a sensor array 100, typically comprising multiple multi-pixel line sensors oriented such that their fields of view are in partially overlapping, mutually skewed arrangement and, therefore, require correction
- the multi-pixel line sensors are typically housed within a camera, such as a CCD camera, having electronic shutters
- a conveyor 102 is arranged to transport an article to be inspected past the sensor array 100 in a transport direction indicated by an arrow 104
- mapping function generator 106 is arranged to receive outputs from the sensor array 100 when a test pattern 108 is being inspected by the sensor array
- the mapping function generator 106 provides an correction output to correction circuitry 1 10 which employs the mapping function generated by mapping function generator 106
- Circuitry 1 10 receives outputs from the sensor array 100 when an article to be inspected, such as a printed circuit board 1 1 1, is being inspected by the sensor array 100
- test pattern 108 is inspected by the sensor array 100 in order whereupon mapping function generator 106 generates the information required to correct the output of circuitry 1 10
- Multiple articles to be inspected may then be inspected and thereafter, intermittently during the inspection of such articles, test pattern 108 may again be inspected by the sensor array 100 to provide updated calibration
- test pattern 108 is preferably inspected about once per month of continuing operation of the inspection system
- Image correction circuitry 1 10 is operative to employ the correction input received from mapping function generator 106 to correct the outputs received from sensor array 100 and provides a corrected sensor array output to segmentation circuitry 1 12 Segmentation circuitry 1 12 provides a segmentation output indication dividing all areas on the image of the article represented by the corrected sensor array output into one of typically two categories For example, in the case of inspection of printed circuit boards, every location on the image represented by the corrected sensor array output is identified by the segmentation output indication as being either laminate or copper
- Image processing circuitry 1 14 is preferably a morphology-based system, but may alternatively be based on a bit map, a net list, or any other suitable input Circuitry 1 14 provides an image processing output which identifies va ⁇ ous features of the image represented by the corrected sensor array output and the locations of such features In the case of printed circuit boards, the features are typically pads, conductor junctions, open ends, and other printed circuit board elements
- the image processing output of circuit 1 14 is supplied to feature registration circuitry 1 16, which maps the coordinate system of the image processing output of circuitry 1 14 onto a feature reference coordinate system, in accordance with information supplied by a reference input source 1 18
- the output of registration circuitry 1 16 and an output of reference input source 1 18 are supplied to feature comparison circuitry 120, which compares the mapped image processing output of circuitry 1 14 with a reference stored in source 1 18 and provides a defect indication which is supplied to a defect indication output generator 122
- Fig 2 is a defect indication which is supplied to a defect indication output generator 122
- the uniformly spaced parallel inclined lines 132 are preferably angled at a small angle ⁇ having a tangent of about 0 05 with respect to the transport direction 104
- the angular o ⁇ entation determinator 134 is preferably nearly rectangular in shape, however one of the edges of the angular o ⁇ entation determinator is preferably angled at a small angle ⁇ having a tangent of about 0 0156 with respect to the transport direction 104
- Fig 3 is a simplified block diagram illustration of the mapping function generator 106 of Fig 1 Outputs 200 representing images of targets on test pattern 108 (Figs 1 and 2) which is being inspected by the sensor array 100 (Fig 1 ) are supplied to the mapping function generator 106 which carries out the following functions Pixel Size and Shape Function Determination for a Single Color 202,
- the parameters thus determined are supplied to a geometric polynomial generator 212 which preferably provides a function which maps the locations viewed by individual elements of sensor array 100 (Fig 1 ) in at least two dimensions
- Fig 4A is a simplified flow chart illustrating operation of the Pixel Size and Shape Function Determination circuitry 202 of Fig 3 and to Fig 4B which illustrates a distortion sought to be overcome by the functionality of circuitry 202
- the apparent dimensions of identical features as seen by camera 124 may vary depending on the position of the feature in the field of view of the camera for a given field of view angle ⁇
- an "on-axis" feature 140 which is directly in front of the camera and in the illustration spans the field of view of camera 124 is perceived to have a width of "d" pixels
- an identical feature 142 when located at the edge of the field of view does not span the field of view of camera 124 and is perceived to have a width "d - ⁇ " pixels
- row 130 of parallel uniformly spaced inclined lines 132 of test pattern target 108 (Fig 2) is viewed by a plurality of cameras 124, of which only one is shown in Fig 4C
- Each camera 124 acquires an image of a part of row 130, an enlarged part of which is shown at reference numeral 150
- the image is preferably acquired in a single color, such as red Alternatively, the image may be acquired in several colors, one of which may be used for the purpose of calculating the size and pixel shape
- the angle ⁇ at which images of the lines 132 are inclined with respect to the transport direction 104, is measured
- the images of the lines are indicated by reference numerals 152
- the separation between each adjacent line 132 in row 130 is fixed and predetermined, and the location of each line 132 in row 130 along an axis 105 that is perpendicular to transport direction 104 is known with respect to an arbitrarily chosen one of
- each line 152 produces a local minimum in graph 160
- the separation of adjacent lines 152 in the camera output may be determined by measuring the distance between adjacent points of inflection before each local minimum 154' indicated the point of inflection corresponding to line location 154
- the X-axis of graph 160 represents the number of each initial individual diode in a linear array of diodes in camera 124, while the Y-axis in graph 160 represents the summation of the intensity L of the image scanned in a direction angled by angle ⁇ with respect to direction 105
- Fig 5A is a simplified flow chart illustrating operation of the Test Pattern Angle Determination circuitry 204 of Fig 3 and to Fig 5B which illustrates the distortion sought to be overcome by the functionality of circuitry 204
- the entire test pattern 108 is normally, in reality, not perfectly aligned with the transport direction 104, but rather is offset therefrom by an angle ⁇
- test pattern target 108 comprising angular o ⁇ entation determinator 134, having edge 182 of known aspect angle ⁇ relative to additional objects in the test pattern, is viewed by camera 124
- the camera 124 acquires an image of angular orientation determinator 134, an enlarged portion of which is shown at reference numeral 184 in
- An aspect angle ⁇ * of an inclined edge 186 of the image of the determinator 134 is calculated by conventional techniques based on measurements carried out on each raster line of the image
- the thus-determined deviation is employed for correction in circuitry 1 10 (Fig
- Fig 6A is a simplified flow chart illustrating operation of the Determination of Relative Orientation of Sensors by Correlating Images of Test Pattern Targets at Edges of the Field of View of Sensors circuitry 206 of Fig 3 and to Fig 6B which illustrates the distortion sought to be overcome by the functionality of circuitry 206 As seen in Fig 6B, rather than being aligned in a single row or
- Fig 7A is a simplified flow chart illustrating operation of the Determination of X Overlap and Y Offset of Sensors by Correlation of Adjacent Images 208 of Fig 3 and to Fig 7B which illustrates the distortion sought to be overcome by the functionality of circuitry 208
- the fields of view 162 of various cameras 124 are seen, in an exaggerated view, to be mutually skewed and shifted in what is referred to herein as the Y direction, being the same as direction 104, and partially overlapping in what is referred to herein as the X direction, being perpendicular to direction 104
- the functionality described here with reference to Figs 7A - 7C deals with the problem of offset of the fields of view of the cameras 124
- An overlap in the fields of view of two adjacent cameras is shown at reference numeral 348
- test pattern target 108 preferably comprising row 300 of LORs 136 as in Fig 6C, is viewed by multiple cameras 124 Each camera 124 acquires an image of part of row 300 which includes the same LORs
- the X overlap and Y offset may be determined by using an image region 253 that is acquired by two adjacent cameras 124
- Two enlarged images of one image region 253 of a LOR 136, as seen by two adjacent cameras 124, are shown in Fig 7C at reference numeral 354
- the two enlarged images of LOR 136 are shown in mutually offset, overlapping relationship, such that the LORs seen in both images are in precise overlapping registration
- the offsets being expressed in pixel units ⁇ x Q v may then be converted to metric units and expressed as by employing the pixel size and shape function as desc ⁇ bed hereinabove with reference to Figs 4A - 4C
- An overlap in the x direction, OV may then be calculated as w-m ⁇ x 0 v, where w is the metric width of each image may be converted to metric units by multiplying ⁇ y
- Fig 8A is a simplified flow chart illustrating operation of the Determination of X and Y Offsets for Multiple Colors circuitry 210 of Fig 3 and to Fig 8B which illustrates a distortion sought to be overcome by the functionality of circuitry 210 As seen in Fig 8B, a three-color CCD camera 380 is shown Camera
- 380 typically includes a multi-pixel line sensor 382 comprising three line sensors such as 384, 386, and 388, with the line sensor being arranged in parallel and each comp ⁇ sing a plurality of single-color sensing diodes 390 arranged linearly
- the diodes 390 of the multi-pixel line sensor 382 may be logically arranged into groupings of three diodes, one from each of line sensors 384, 386, and 388, with each diode sensing a different color Three such groupings 392, 394, and 396 are shown, at the center and both edges of camera 380
- Camera 380 is shown viewing elements 398 of a target 400 moving in direction 104
- the image acquired by the multi-pixel line sensor 382 is typically stored in three buffers 402, 404, and 406, with each buffer corresponding to a particular color, such as red, green, and blue respectively
- a combined view of buffers 402, 404, and 406 is shown at reference numeral 408
- Combined buffer 408 shows the acquired images 399, 401 , and 403 of elements 398
- the images 399, 401, and 403 of combined buffer 408 demonstrate the pixel size and shape distortion in the X direction due to chromatic aberration, as shown at 410, as well as the Y direction displacement, as shown at 412, due to the physical separation of the R, G and B line sensors
- test pattern target 108 preferably comprising row 300 of LORs 136 as seen in Figs 6C and 7C, is viewed by multiple cameras 124 Each camera 124 preferably acquires a multicolored image at either edge of the camera, with one edge of camera 124 acquiring an image of one LOR 136, and the other edge acquiring an image of an adjacent LOR 136
- Each color of each multicolored image acquired at each edge of one of the cameras 124 is preferably related to separately
- a single color, such as red may be selected as a reference color to which the other two color components of the multicolored image are compared
- the red and green components of image region 253 at one edge of the field of view of CAM 1 are shown enlarged at reference numeral 360
- the two enlarged images of LOR 136 are shown in mutually offset, overlapping relationship, such that the LORs seen in both images are in precise overlapping registration It is noted that due to this overlapping relationship between the images a y offset, ⁇ yco L .
- Figs 9A and 9B which, taken together, are simplified flow charts illustrating operation of the geometric polynomial generator 212 of Fig 3
- a cubic function may be constructed to determine the positions of the diodes of cameras 124 in space using the outputs of 202 - 210 described hereinabove with reference to Figs 4A - 8C
- Fig 9A the X-overlap, OV x , as determined in 208 (Fig 3), is used to find a0[n] first for one color, such as red, of each camera n in a row of cameras, such as cameras designated CAM 1 , CAM 2, and CAM 3 in Fig 7C
- the X- polynomials may then be derived for the other colors, such as blue and green, based on the X-polynomial for the first color as follows
- ND is the number of diodes in the preceding camera n-1
- S n . ⁇ (ND) is the value of the pixel size and shape function output determined in 202 as evaluated for the last pixel of camera n-1
- w is the width of the image in metric units containing the LOR
- OVx is the measured overlap in the X-direction as determined m 206
- the Y-polynomtal is determined initially for one color, such as red, of each camera n in a row of cameras, such as cameras designated CAM 1, CAM 2, and CAM 3
- the Y-polynomials may then be derived for the other colors, such as blue and green, based on the Y-polynomial for the first color
- Fig 10A is a simplified illustration of multiple cameras 500 and 502 acquiring an image of a target 504 under ideal conditions
- Fig 10B which is a simplified illustration of image buffers for storing the acquired image of target 504
- Cameras 500 and 502 are typically in fixed positions, each having a static field-of-view, and are arranged such that target 504 passes through the fields-of-view of cameras 500 and 502 in the direction of motion 104
- Cameras 500 and 502 each acquire an image of target 504 one image line at a time by employing a multi-pixel line sensor as is desc ⁇ bed hereinabove
- Each diode of the multi-pixel line sensor acquires a single-pixel image of a specific location on target 504, and the pixels acquired by each diode collectively form an image line
- An image line portion 512 is shown comp ⁇ sing a plurality of pixels 514
- the field- of-view of cameras 500 and 502 "moves" in the direction designated by arrows 508, and thus the image lines are acquired in the direction of 508 as well
- camera 500 begins acquiring an image at t 0 to yield an image line 516 shown in dashed lines
- Image line portion 512 is acquired at time index t s , shown intersecting a portion of a target element 518
- Target 504 continues to move in the direction of arrow 506, and the acquired image lines approach t] as is shown in dashed lines by an image line portion 520
- Fig 10 A image line portion 512 associated with camera 500 is aligned in a single row with an image line portion 522 associated with camera 502 at time index t , with image line portions 512 and 522 meeting at a boundary line 524
- the image lines scanned by cameras 500 and 502 are typically stored in buffers, such as buffers 530 and 532 of Fig 10B Since the scan lines of both buffers are aligned in a single row for each corresponding time index, buffers 530 and 532 may be combined to form a non-distorted composite image of target element 518, a portion of which is shown as 534
- Figs 1 1 A and 1 IB illustrate the effect cameras 500 and 502 have when acquiring an image of target 504 under less than ideal conditions in contrast to Figs 10A and 10B, specifically when the fields of view of both cameras are mutually skewed and are overlapping
- Figs 1 1 A and 1 IB are intentionally provided as an oversimplified illustration of some difficulties encountered, and are merely intended to review what was described in greater detail hereinabove and are not intended to supersede the descriptions of Figs 1 - 9B
- the image lines acquired are shown not aligned in a single row for a given time index t , such as is shown with reference to buffers 560 and 562 and image line portions 564 and 566 of Fig 1 I B
- Combining buffers 560 and 562 would produce a distorted composite image 568 of target element 518, as is shown in Fig 1 IB
- simply combining the buffer images would neither correct for image overlap, discussed hereinabove with reference to Figs 7A - 7C, nor for the viewing angle distortion, discussed hereinabove with reference to Figs 4A - 4C
- a FIFO buffer such as a FIFO buffer 600 of Fig 12, may be defined by defining a window having a height expressed in a fixed number of image buffer rows, such as 40, beginning with the first row of pixels acquired This window is then typically advanced to a new position one row at a time as each new row of pixels is acquired Alternatively, an image buffer may initially be filled with row of pixels, at which point the FIFO buffer window may be defined as a subset of the image buffer rows and advanced along the image buffer in the manner just described Before the X and Y polynomials determined with reference to Figs 9A and 9B can be used for correcting the image they may be translated into another type of polynomial referred to herein as a "diode compensating polynomial" This polynomial maps the relationship between a diode and a pixel position of a corrected image constructed using a pixel size chosen by the user The diode compensating polynomials Q x (d) and Q y (d) may be derived from
- Q y (d) Py(d)/p
- p is the pixel size chosen by the user and is expressed in the same metric units as Px(d)
- the pixel size p chosen by the user must be a multiple of the minimum measurable distance of travel in the scan direction 104, typically one pulse unit of a drum encoder
- the diode compensating polynomial Q may be additionally be adjusted for the shift introduced by different color component accumulation times as is described in greater detail hereinbelow with reference to Fig 17
- Each pixel or grid point in the FIFO buffer represents the sampling of a corresponding target acquired by a diode
- a process of "resampling" is used whereby calculations are performed to determine a gray level g at an arbitrary position "between" grid points of the buffer grid
- the four gray levels g may be selected from four contiguous grid points of the FIFO buffer These four points are referred to herein as a "quartet"
- Resampling may be performed in two stages corresponding to the X and Y directions described hereinabove
- the stages are referred to herein as X- resamphng and Y-resamphng Y-resamphng is now described in greater detail with reference to Fig 12
- Two FIFO buffers 600 and 602 are shown corresponding to two cameras
- the set of all pixels 604 in the FIFO buffers scanned by a diode d may be referred to as diode d's "gray level column," such as a column 606
- Virtual scan lines 608 and 610 are shown to indicate the correction angles needed for each buffer to compensate for the misalignment angles of each corresponding camera
- a quartet 612 is shown as a group of four pixels within the gray level column 606 closest to
- the quartet index q(d) denotes the first grid point belonging to the quartet of pixels that he within the diode's gray level column
- the index q(d) for each quartet is preferably stored in a quartet look-up table in a position corresponding to the diode d
- these four coefficients cl , c2, c3, and c4 are preferably encoded in such a manner that when decoded and summed the summed value equals 1 0, although the accuracy of any single coefficient may be diminished
- the encoded values are preferably stored in a coefficients look-up table in a position corresponding to d
- a "blending region" 626 of a predefined number of pixels B is defined within the overlap region 624 between cameras 1 and 2 where corresponding pixels from both cameras within the blending region are blended to yield a single pixel value which is then used to form the combined image 630
- the blending region preferably begins after allowing for a pixel margin 628 of a predefined number of pixels M, typically 20 pixels, in order to accommodate lower-quality output often encountered at the ends of a diode array
- the two diode resampling polynomials ON (d) and Q (2) (d) of the two adjacent cameras may be used to determine the amount of overlap between the cameras
- the following steps may be performed 1 Defining Q (2) (l ) to refer to the pixel position r of the first pixel of X and Y-resampled output of camera 2 (X-resampling is described in greater detail hereinbelow with reference to Fig 15)
- the leading edge of the blending region may be determined by adding the predetermined number of pixels defined for the pixel margin desc ⁇ bed above
- X-resampling is now described in greater detail with reference to Fig 15 Due to optical distortions, and in order to accommodate a user-defined pixel size, an output 640 of the Y-resamphng must be resampled in the X direction, thereby creating a X-corrected image row 642 with pixels having the desired pixel size
- Each pixel position r on the X-corrected image row corresponds to a position d(r) on the diode array d(r) may be calculated using the diode resampling polynomial Q x (d) This involves finding the inverse function Q '(r) of Q x (d) This inverse function allows the mapping of the pixel position r on the X-corrected image row to a corresponding position d(r)on the diode array It is appreciated that this position might not correspond to a integer position of a specific diode, but rather may be expressed in fractional terms along the diode array Once the dio
- Fig 16 illustrates the steps to be performed for each pixel in the X-corrected image row as follows 1 Assigning an index r p to correspond to the first pixel 650 in an
- Stepping index r p through each pixel position in the X- corrected image row 652 corresponding to an overlap region 654 to the end of the field of view 656 of the current camera, CAM 1 , and find the diode position d p (1) such that rp (1 ) Qx(d p (1 ) ) Since Qx is a monotonically increasing function, d p advances as r p advances When d p reaches the end 656 of the field of view of CAM 1, r p is returned to the pixel posi'ion corresponding to the start of the overlap region 658 of the next camera, CAM 2, assigning to r p the value of the diode compensating polynomial evaluated for the first diode of camera CAM 2, I e r p ⁇ — Q (2) ( 1 ) 3 Stepping index r p through each pixel position in the X- corrected image row 652 for CAM 2, finding the diode position d p ( ) such that r P (2) -
- Steps 2 and 3 are performed for each subsequent pair of cameras
- Finding the diode position d p may either be done through a one-time inversion of the function Qx(d p ) or through a numerical solution
- the convolution coefficients cl through c4 may be calculated based on ⁇ using the formulae described above for Y- resamp ng
- Image correction employing Y-Resamphng, X-Resamphng, and Overlap correction are preferably performed by circuitry 1 10 of Fig 1, now summarized hereinbelow
- the Y-Resampled output is processed further to correct pixel shape and size
- the gray level is processed one pixel row at time
- the X-quartet index is retrieved from the X-quartet look-up table and the four convolution coefficients Ci through c 4 are retrieved from the X-coefficients look-up table
- the four gray level values g through g - of the corresponding X-quartet may then be retrieved from the Y- corrected gray level buffer
- each color component of the multi-line sensor array begins to accumulate the charge corresponding to its respective color
- the exposure of each color component of the multi-line sensor array is then varied by closing the electronic shutters of each color component at different times
- the center of the acquired pixel for each color component may be different than the geometric center of the pixel
- an "accumulation shift" ⁇ y aCL is introduced that may be corrected by subtracting the center of the acquired pixel from the geometric center of the pixel for each color component by the formula
- This accumulation shift is preferably determined during acquisition of the test target, and is used to adjust the bO component of the Y-polynomial P y
- the diode compensating polynomial Q> described in Fig 9B may also be adjusted for the accumulation shift according to the different exposure times chosen for the various color components
- a multi-line sensor array 700 is shown acquiring three pixels 702, 704, and 706, each pixel being acquired by a different line sensor 708, 710, and 712, with each line sensor comprising a plurality of single-color sensing diodes Due to the different acquisition times of each of each line sensor, the relative areas of each of the three pixels acquired vary, as is shown by accumulation areas 714, 716, and 718
- a geometric center may be defined for each of the three pixels at 720, 722, and 724
- the center of each accumulation area may be defined at 726, 728, and 730
- the distances between the centers of each accumulation area and its corresponding geometric center 732, 734, and 736 represent the accumulation shift for each color component and may be used to correct for the overlap in the Y-direction 104 as described above It is appreciated that various features of the invention which are, for clarity, described in the contexts of separate embodiments may also be provided in combination in a
Abstract
Description
Claims
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
AU53840/99A AU5384099A (en) | 1998-08-25 | 1999-08-19 | Method and apparatus for inspection of printed circuit boards |
EP99939581A EP1108329A1 (en) | 1998-08-25 | 1999-08-19 | Method and apparatus for inspection of printed circuit boards |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
IL12592998A IL125929A (en) | 1998-08-25 | 1998-08-25 | Method and apparatus for inspection of printed circuit boards |
IL125929 | 1998-08-25 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2000011873A1 true WO2000011873A1 (en) | 2000-03-02 |
Family
ID=11071888
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/IL1999/000450 WO2000011873A1 (en) | 1998-08-25 | 1999-08-19 | Method and apparatus for inspection of printed circuit boards |
Country Status (5)
Country | Link |
---|---|
EP (1) | EP1108329A1 (en) |
CN (1) | CN1314049A (en) |
AU (1) | AU5384099A (en) |
IL (2) | IL125929A (en) |
WO (1) | WO2000011873A1 (en) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE10208289C1 (en) * | 2002-02-26 | 2003-02-27 | Koenig & Bauer Ag | Electronic image sensor with read out of signals from individual sensors so that overlapping image regions are only read once |
CN1306244C (en) * | 2005-06-16 | 2007-03-21 | 姚晓栋 | On-the-spot printing circuit board test based on digital image |
CN102914543A (en) * | 2011-08-03 | 2013-02-06 | 浙江中茂科技有限公司 | Article detection device of three-dimensional stereo image |
CN107860773B (en) * | 2017-11-06 | 2021-08-03 | 凌云光技术股份有限公司 | Automatic optical detection system for PCB and correction method thereof |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5099522A (en) * | 1989-05-29 | 1992-03-24 | Rohm Co., Ltd. | Method and apparatus for performing head-tail discrimination of electronic chip components |
US5298989A (en) * | 1990-03-12 | 1994-03-29 | Fujitsu Limited | Method of and apparatus for multi-image inspection of bonding wire |
US5686994A (en) * | 1993-06-25 | 1997-11-11 | Matsushita Electric Industrial Co., Ltd. | Appearance inspection apparatus and appearance inspection method of electronic components |
-
1998
- 1998-08-25 IL IL12592998A patent/IL125929A/en not_active IP Right Cessation
-
1999
- 1999-08-19 CN CN99809996.1A patent/CN1314049A/en active Pending
- 1999-08-19 WO PCT/IL1999/000450 patent/WO2000011873A1/en not_active Application Discontinuation
- 1999-08-19 EP EP99939581A patent/EP1108329A1/en not_active Withdrawn
- 1999-08-19 AU AU53840/99A patent/AU5384099A/en not_active Abandoned
-
2002
- 2002-01-17 IL IL14772302A patent/IL147723A0/en unknown
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5099522A (en) * | 1989-05-29 | 1992-03-24 | Rohm Co., Ltd. | Method and apparatus for performing head-tail discrimination of electronic chip components |
US5298989A (en) * | 1990-03-12 | 1994-03-29 | Fujitsu Limited | Method of and apparatus for multi-image inspection of bonding wire |
US5686994A (en) * | 1993-06-25 | 1997-11-11 | Matsushita Electric Industrial Co., Ltd. | Appearance inspection apparatus and appearance inspection method of electronic components |
Also Published As
Publication number | Publication date |
---|---|
IL147723A0 (en) | 2002-08-14 |
IL125929A (en) | 2002-03-10 |
AU5384099A (en) | 2000-03-14 |
IL125929A0 (en) | 1999-04-11 |
CN1314049A (en) | 2001-09-19 |
EP1108329A1 (en) | 2001-06-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US4311914A (en) | Process for assessing the quality of a printed product | |
JP3525964B2 (en) | 3D shape measurement method for objects | |
US6876763B2 (en) | Image resolution improvement using a color mosaic sensor | |
US7315643B2 (en) | Three-dimensional shape measurement technique | |
JP3937024B2 (en) | Detection of misalignment, pattern rotation, distortion, and misalignment using moiré fringes | |
WO2008011888A1 (en) | Autostereoscopic system | |
JP4966413B2 (en) | Method and apparatus for coping with chromatic aberration and purple fringing | |
JP2004109106A (en) | Method and apparatus for inspecting surface defect | |
CN106017313B (en) | Edge detection deviation correction value calculation method, edge detection deviation correction method and device | |
JP2009522561A (en) | Method and system for optical inspection of periodic structures | |
US20060018533A1 (en) | Segmentation technique of an image | |
JP3924796B2 (en) | Pattern position measuring method and measuring apparatus | |
US5852671A (en) | Method for reconstructing a curved surface of a subject and an apparatus for reconstructing a curved surface of a subject | |
US6813392B2 (en) | Method and apparatus for aligning multiple scans of the same area of a medium using mathematical correlation | |
EP1108329A1 (en) | Method and apparatus for inspection of printed circuit boards | |
EP1692869A2 (en) | Inspection apparatus and method | |
JP2004077290A (en) | Apparatus and method for measuring three-dimensional shape | |
CN111080716B (en) | Camera calibration target based on color coding phase shift stripes and calibration point extraction method | |
JP3433331B2 (en) | Image inspection method and apparatus | |
CN105894068B (en) | FPAR card design and rapid identification and positioning method | |
KR100303181B1 (en) | Calibration method of high resolution photographing equipment using multiple imaging device | |
JP2961140B2 (en) | Image processing method | |
JP4985213B2 (en) | Three-dimensional shape measuring method, apparatus and program | |
JP3563797B2 (en) | Image inspection method and apparatus | |
JP2675167B2 (en) | Pattern recognition method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 99809996.1 Country of ref document: CN |
|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AE AL AM AT AT AU AZ BA BB BG BR BY CA CH CN CR CU CZ CZ DE DE DK DK DM EE EE ES FI FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SK SL TJ TM TR TT UA UG US UZ VN YU ZA ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): GH GM KE LS MW SD SL SZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
WWE | Wipo information: entry into national phase |
Ref document number: 1999939581 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 1999939581 Country of ref document: EP |
|
REG | Reference to national code |
Ref country code: DE Ref legal event code: 8642 |
|
WWW | Wipo information: withdrawn in national office |
Ref document number: 1999939581 Country of ref document: EP |