GB2262339A - Method of inspecting the surface of a workpiece - Google Patents

Method of inspecting the surface of a workpiece Download PDF

Info

Publication number
GB2262339A
GB2262339A GB9225600A GB9225600A GB2262339A GB 2262339 A GB2262339 A GB 2262339A GB 9225600 A GB9225600 A GB 9225600A GB 9225600 A GB9225600 A GB 9225600A GB 2262339 A GB2262339 A GB 2262339A
Authority
GB
United Kingdom
Prior art keywords
image
workpiece
image data
light
workpiece surface
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
GB9225600A
Other versions
GB9225600D0 (en
GB2262339B (en
Inventor
Norihide Kato
Tomohide Shimizu
Kenichiro Mori
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Honda Motor Co Ltd
Original Assignee
Honda Motor Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from JP3329384A external-priority patent/JPH05164700A/en
Priority claimed from JP3329383A external-priority patent/JP2941107B2/en
Priority claimed from JP3333404A external-priority patent/JPH05164703A/en
Priority claimed from JP3337263A external-priority patent/JP2904983B2/en
Application filed by Honda Motor Co Ltd filed Critical Honda Motor Co Ltd
Publication of GB9225600D0 publication Critical patent/GB9225600D0/en
Publication of GB2262339A publication Critical patent/GB2262339A/en
Application granted granted Critical
Publication of GB2262339B publication Critical patent/GB2262339B/en
Anticipated expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0004Industrial image inspection
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/84Systems specially adapted for particular applications
    • G01N21/88Investigating the presence of flaws or contamination
    • G01N21/95Investigating the presence of flaws or contamination characterised by the material or shape of the object to be examined
    • G01N21/9515Objects of complex shape, e.g. examined with use of a surface follower device
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20004Adaptive image processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20036Morphological image processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30108Industrial image inspection
    • G06T2207/30156Vehicle coating

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Chemical & Material Sciences (AREA)
  • Biochemistry (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Quality & Reliability (AREA)
  • Analytical Chemistry (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Immunology (AREA)
  • Pathology (AREA)
  • Image Processing (AREA)
  • Investigating Materials By The Use Of Optical Means Adapted For Particular Applications (AREA)
  • Length Measuring Devices By Optical Means (AREA)

Abstract

A method of inspecting a surface of a workpiece uses apparatus provided with a projecting device for illuminating the workpiece surface with detection light, a detecting device for receiving and detecting the light reflected from the workpiece surface, and a condensing optical system for causing the reflected light to converge on a light-receiving surface of the detecting device. According to the method, the surface inspecting apparatus is successively displaced along the workpiece surface and the reflected light is detected by the detecting device. Further, image data obtained from the detecting device is held and binary digitized. It is determined whether or not the workpiece surface to be inspected is planar. If it is not planar, then the digitized image is processed for dilation and erosion. Thus, when the workpiece surface is curved, faulty or deflective spots can be reliably detected. If the workpiece surface is planar, the time required to process an image can be shortened. The use of windows to define an effective image region and combining windows from several images to increase the area to be processed is also disclosed. <IMAGE>

Description

2262339 METHOD OF INSPECTING THE SURFACE OF A WORKPIECE The present
invention relates to a workpiece surface inspecting method for detecting a defect in a surface such as a coated surface of a workpiece, and the degree of smoothness of the surface.
In the P,-ior art a surface is inspected in the following manner. As shown in FIG. 1A, a workpiece surface 6 is irradiated via a convex lens 4 with light emitted from a projecting means 2 mounted to an arm or the like of an unillustrated robot. Then, the light is caused to converge on a detecting means 8 such as a CCD camera or the like. Further, an image 10a (see FIG. IB) which has been detected by the detecting means 8, is subjected to image processing. Thereafter, the state of coating of the workpiece surface 6 is examined to thereby make a decision as to whether or not dust is being applied on the workpiece surface 6 and flaws exist thereon. In this case, an inspection region at which the workpiece surface 6 can be inspected at a time, can be enlarged by inspecting the workpiece surface 6 using the converged light.
In this case, the detected light satisfactorily falls on a lightreceiving surface of the detecting 1 means 8 if the workpiece surface 6 is planar. When, however, the workpiece surface 6 is curved as illustrated in FIG. 2A, the detected light is diffused from the workpiece surface 6. Therefore, a bright spot or area 12 (which is the effective inspection surface) and shadow areas 14 (which are unusable as an object to be inspected), are developed in a detected image 10b (see FIG. 2B). Further, a blurred portion is developed at the respective boundaries between the bright area 12 and each of the shadow areas 14. That is, small black spots or points (hereinafter called "isolated spots or points") tend to appear on the bright area 12 side owing to the influence of the state of coating of the workpiece surface 6. Thus, when subsequent image processing is effected, the isolated points cannot be judged as being isolated points produced by the adhesion of dust, for example. Accordingly, each of the isolated points tends to be misjudged as being defects in coating. Therefore, there has been proposed a method of processing detected image data for dilation and erosion to thereby remove the isolated points from the image data.
When, however, all the detected image data are processed for dilation and erosion as described above, the time required to process the image is prolonged and the total inspection time is increased.
When the image processing is made on the entire 2 surface of the detected image 10b, the image processing is performed even on the shadow area 14 in addition to the bright area 12, which is the effective inspection surface. Accordingly, the time required to inspect the workpiece surface is wasted and hence the entire inspection time is lengthened.
Further, such a prior art has been disclosed in Japanese Laid-Open Patent Publication No. 1-180438, for example. According to the disclosure, a brightness/darkness-striped image projected on a surface to be inspected, is photographically recorded by an exposure means as a level signal indicative of strength and weakness of light. Then, the amount of light emitted from a light source is adjusted according to the strength of the level signal. The level signal can be detected within a predetermined range by adjusting the amount of the light emitted, irrespective of the difference in colors applied on the workpiece surface to be inspected and of any deterioration in the light source. Accordingly, any defect in the workpiece surface to be inspected can be reliably detected without a change in a predetermined defect criterion.
When the surface to be inspected is planar, defects in the surface can be satisfactorily detected once the amount of light is adjusted. The above prior art is, however, accompanied by the problem that since
3 i the amount of light reflected by the surface to be inspected varies according to the state of exposure when the surface is curved, it is necessary to determine whether or not the level signal indicative of the strength and weakness of the light falls within a predetermined level range each time the surface to be inspected is photographically recorded and to repeatedly adjust the amount of the light accordingly. The total time required to inspect the surface is increased thereby.
It is a principal aim of the present invention to provide a method of inspecting the surface of a workpiece, which is capable of reducing the time required to process an image and shortening the entire inspecting time by selectively processing detected image data for dilation and erosion.
It is another aim of the present invention to provide a method of inspecting the surface of a workpiece, which can reduce the time required to process an image and shorten the entire inspecting time by performing image processing only on an effective inspection surface of a detected image.
It is a further aim of the present invention to provide a method of inspecting the surface of a workpiece, which can make it unnecessary to adjust the 4 I amount of light irrespective of the shape of a surface to be inspected and suitably detect defective points from respective photographically- recorded images, thereby making it possible to shorten the entire inspection time.
The present invention provides a method of inspecting a surface of a workpiece by using a surface inspecting apparatus provided with projecting means for illuminating the workpiece surface to be inspected with detection light, detecting means for receiving and detecting the light reflected from the workpiece surface, and a condensing optical system for causing the reflected light to converge on a light-receiving surface of the detecting means, comprising the following steps of: successively displacing the surface inspecting apparatus along the workpiece surface, receiving and detecting the reflected light by the detecting means and holding image data obtained from the detecting means, binary digitizing the image data, determining whether or not the workpiece surface to be inspected is planar and processing the digitized image data for dilation and erosion when the workpiece surface to be inspected is non- pianar.
4 The present invention also provides a method of inspecting a surface of a workpiece by using a surface inspecting apparatus provided with projecting means for irradiating the workpiece surface to be inspected with detection light, detecting means for receiving and detecting the light reflected from the workpiece surface, and a condensing optical system for causing the reflected light to converge on a light-receiving surface of the detecting means,comprising the following steps of: selecting a preset window for extracting, based on the shape of the workpiece surface to be inspected, an effective image to be inspected from an image obtained by the detecting means, and performing image processing on an image included in an effective inspection region, the image having been extracted by using the selected window.
The present invention further provides a method of inspecting a surface of a workpiece by using a surface inspecting apparatus provided -with projecting means for illuminating the workpiece surface to be inspected with detection light, detecting means for receiving and detecting the light reflected from the workpiece surface, and a condensing optical system for causing the reflected light to converge on a light-receiving surface of the detecting means, comprising the following steps of: binary 6 digitizing an image obtained by the detecting means, producing a histogram graphically representing a relationship between bright areas and shadow areas from the digitized image, setting a window for extracting an effective image to be inspected, based on the histogram, extracting an effective image to be inspected from the digitized image by using the set window, and performing image processing on the extracted image.
e present -,jrEher provides a method of inspecting a surface of a workpiece by using a surface inspecting apparatus provided with projecting means for illuminating the workpiece surface to be inspected with detection light, detecting means for receiving and detecting the light reflected from the workpiece surface, and a condensing optical system for causing the reflected light to converge on a light-receiving surface of the detecting means, comprising the following steps: a first step for successively displacing the surface inspecting apparatus along the workpiece surface, receiving and detecting the reflected light by the detecting means and by-level digitizing a plurality of images obtained by the detecting means, a second step for extracting image data corresponding to an effective Lnspection surface of the workpiece irradiated 7 i with the detection light from the plurality of digitized image data and for holding the extracted image data in a non-superimposed state, and a third step for subjecting the held image data to given image processing.
Preferablv, a method of inspecting a surface of a workpiece, further comprises a fourth step for teaching the surface inspecting apparatus in advance so as to avoid the superimposition of the plurality of image data on one another.
Preferably, the second step includes a step for setting a window used to define a region corresponding to the effective inspection surface, based on the extracted digitized image data and a step for holding image data corresponding to the effective inspection surface in a non-superimposed state, the image data having been extracted by the window.
The present invention further provides a method of inspecting a surface of a workpiece by using a surface inspecting apparatus provided with projecting means for illuminating the workpiece surface to be inspected with detection light, detecting means for receiving and-detecting the 8 1 light reflected from the workpiece surface, and a condensing optical system for causing the reflected light to converge on a light-receiving surface of the detecting means, comprising the following steps of: successively displacing the surface inspecting apparatus along the workpiece surface, receiving and detecting the reflected light by the detecting means and holding images obtained by the detecting means as tone image data every pixels, creating a histogram graphically representing the number of pixels corresponding to respective tone levels based on the tone image data, setting a threshold value represented in a digitized level between the maximum value of a tone level in the histogram as seen on the bright area side and the maximum value of a tone level in the histogram as seen on the shadow area side, converting the tone image data into digitized image data based on the threshold value, and subjecting the digitized image data to image processing.
Embodiments of the present invention will now be described by way of Example only with reference to --he accompanying drawin;s, jn-which:- 9 FIG. 1A is a view showing the manner in which an image of the surface of a workpiece is photographically recorded based on a conventional workpiece surface inspecting method; FIG. 1B is a view for describing the image of the surface to be inspected, which has been photographically recorded under the condition of FIG. 1A; FIG. 2A is a view showing the manner in which an image of the curved surface of a workpiece is photo graphically recorded based on another conventional workpiece surface inspecting method; FIG. 2B is a view for describing the image on the surface to be inspected, which has been photographically recorded under the condition of FIG. 2A; FIG. 3 is a schematic view illustrating a workpiece surface inspecting apparatus for executing a workpiece surface inspecting method according to one embodiment of the present invention; FIG. 4 is a schematic view showing an image processing device according to the present invention; FIG. 5 is an overall flowchart for describing the workpiece surface inspecting method of the present invention; FIG. 6 is a fragmentary flowchart for describing the setting of a threshold value represented in a digitized level; FIG. 7A is a view illustrating the manner in 1 which an image on the surface of a workpiece is photographically recorded; FIG. 7B is a view for describing the image on the surface to be inspected, which has been photographically recorded under the condition of FIG. 7A; FIG. 8 is a histogram graphically representing brightness based on the input image; FIG. 9A is a view illustrating the manner in which an image on the surface of a workpiece is photographically recorded based on a workpiece surface inspecting method according to another embodiment of the present invention; FIG. 9B is a view for describing the image on the surface to be inspected, which has been photographically recorded under the condition of FIG. 9A; is a histogram graphically representing brightness based on the input image; FIG. 11 is a fragmentary flowchart for describing image processing based on the workpiece surface inspecting method shown in FIG. 9A; FIG. 12 is a view for describing images on the surface of the workpiece, which has been inspected based on the workpiece surface inspecting method shown in FIG. 9A; FIG. 13 is a view for describing windows which have been set based on the workpiece surface inspect ing method shown in FIG. 9A; 11 i FIG. 14 is a fragmentary flowchart for describing the workpiece surface inspecting method shown in FIG. 9A; FIG. 15 is another fragmentary flowchart for describing the workpiece surface inspecting method shown in FIG. 9A; FIG. 16A is a view for describing the manner in which a histogram with respect to digitized image data is set based on the workpiece surface inspecting method shown in FIG. 9A; FIG. 16B is a view for describing a window set based on the histogram shown in FIG. 16A; FIG. 17A is a view for describing the manner of setting of another histogram with respect to digitized image data by the workpiece surface inspecting method depicted in FIG. 9A; FIG. 17B is a view for describing a window set based on the histogram shown in FIG. 17A; FIG. 18A is a view for describing the manner of setting of a further histogam, with respect to digitized image data by the workpiece surface inspecting method shown in Fig. 9A; FIG. 18B is a view for describing a window set based on the histogram shown in Fig. 18A.
FIG. 19A is a view for describing the manner of setting of a still further histogram with respect to digitized image data by the workpiece surface inspecting method depicted in FIG. 9A; FIG. 19B is a view for describing a window set 12 based on the histogram shown in FIG. 19A; FIG. 20 is a fragmentary schematic view for describing the combination of images by the workpiece surface inspecting method shown in FIG. 9A; FIG. 21 is a flowchart view for describing the combination of images by the workpiece surface inspecting method shown in FIG. 9A; FIG. 22 is a view for describing a dilation and erosion process based on the workpiece surface inspecting method shown in FIG. 9A; FIG. 23 is a view for describing another dilation and erosion face inspecting method depicted in FIG. 9A; and FIG. 24 is a view for describing a further process based on the workpiece sur- dilation and erosion process based on the workpiece surface inspecting method shown in FIG. 9A.
One embodiment of the present invention will be described based on a workpiece surface inspecting method using a workpiece surface inspecting apparatus employed in a vehicle production line.
As shown in FIG. 3, a workpiece surface inspecting apparatus 20 is mounted on a robot 22. A light source 24 is fixed to an arm 26 of the robot 22. A condensing optical system 30 is fixed to a wrist 34 and is electrically connected to the light source 24 by an optical fiber bundle 28. Optical system 30 comprises a Fresnel lens or the like, and a CCD camera 32 which serves as a detecting means. The CCD camera 32 is electrically connected to an image processing device 36 for processing an image of the surface 35 of a workpiece, for example. The image processing device 36 is electrically connected to a robot controller 38. In the image processing device 36 as shown in FIG. 4, image data inputted from the CCD camera 32 is converted into a digital signal by an A/D converter 40. The converted digital signal is supplied to an image tint memory 42 as halftone or tone image data. Each of the tone image data stored in the image tint memory 42 is converted into binary digitized image data by an image processor 44. The converted image data represented in binary form is inputted to a digitized image memory 46. The tone image data stored in the image tint memory 42 and the digitized image data stored in the digitized image memory 46 are displayed on a monitor TV 50 through a D/A converter 48 as needed. The image processor 44 is electrically connected with a microprocessor 52, a memory 54, an 14 external input/output interface I/F 56 via a bus 58. The external input/output I/F 56 is electrically connected to the robot controller 38 and an external computer 59 for controlling the result of processing.
Next, the operations of the workpiece surface inspecting apparatus 20 and the image processing device 36, which are constructed as described above, will be summarized with reference to a flowchart shown in FIG. 5.
The image processing device 36 and the robot controller 38 are first initialized (Step S2). An image processing pattern to be described later is then inputted to the image processing device 36 from the robot controller 38 (Step S4). After the above setting has been completed, the robot 22 is activated to illuminate or irradiate a predetermined range of the workpiece surface 35, a block, with inspection light emitted from the light source 24 by the optical fiber bundle 28 and the condensing optical system 30. The light reflected from the surface 35 falls on the CCD camera 32, so that image data is inputted to the image processing device 36 (Step S6).
The image data inputted to the image processing device 36 is converted into tone image data which serves as a digital signal by the A/D converter 40. The tone image data thus converted is inputted to the image tint memory 42. The image processor 44 digi- tizes an image so as to be expressed in a bright area and a dark or shadow area, based on a histogram graphically representing the tone image data to thereby set up a digitized level used as an identification reference (Step S8).
Predetermined image processing is effected below based on processing patterns inputted from the robot controller 38 (Step S10).
When a first processing pattern is input from the robot controller 38, i. e., when the workpiece surface 35 to be inspected is planar and the tone image data is represented in binary form, the tone image data is converted into digitized image data based on the digitized level, and the so-converted image data are stored in the digitized image memory (Step S12). Then, the shadow area in the digitized image data, which exist within the bright area therein, is extracted as an isolated spot (Step S14). Based on the size of the isolated spot, it is determined and selected whether or not the isolated spot indicates dust or fine particles (Step S16). The number and sizes of fine particles which exist in one block of the processing pattern, are transferred via the external input/output I/F 56 to the computer 59, where they are stored (Step S18). The above processing is effected with respect to-all of the blocks representing the workpiece surface 35 (Step S20).
16 After the image processing on all of the blocks has been completed, the computer 59 outputs the result of inspection (Step S22).
When second and third processing patterns are input from the robot controller 38, i.e., when the workpiece surface 35 is curved and a window is set to an image taken in by the CCD camera 32, a window having a predetermined processing range is selected and a portion corresponding to a shadow area of the resultant image is deleted for the purposes of further image processing (Step S24). Alternatively, a window is determined by computation based on the distribution of the highlight of the image after the image has been digitized. Then, the shadow area or spot is deleted from the object of the image processing in a manner similar to Step S24 (Step S26).
When fourth and fifth processing patterns are input from the robot controller 38, i.e., when the workpiece surface 35 is curved and a process for combining images is selected, only the bright area is taken out from the bright area and the shadow area of a digitized image. Then, images are combined n times until image data corresponding to one frame is stored (Steps S28, S30). Alternatively, a desired window is -ion in a manner similar to Step determined by computat S26, and images having only the bright areas, which are obtained using the determined window, are stored 17 one frame in the digitized image memory 46 and combined into one so as not to overlap with each other (Steps S32, S30). To prevent dust or the like which exists at the boundary between the bright area and the shadow area from being misdetected, the image thus obtained is subjected to a dilation and erosion process for removing fine particles (Step S34). Thereafter, a dust detecting process, executed in Step 16, and steps subsequent to Step S16, are effected in the same manner as the first processing pattern. By performing the above processing on the image thus combined, the image processing can be collectively effected after the highlights of a plurality of images have been brought into one image, thereby enabling an improvement in image processing speed.
Now, Step S8 for executing the setting of the threshold value represented in the form of the digitized level, will be described in detail with reference to FIGS. 6 through 10.
First of all, an image on the workpiece surface 35 is inputted to the image processing device 36 from the CCD camera 32. The image is then supplied to the image tint memory 42 as tone image data. Thereafter, the image processor 44 produces a histogram graphically representing the number of pixels with respect to each brightness or luminance, based on the tone image data.
18 If the workpiece surface 35 is curved at this time, inspection light projected or emitted from the condensing optical system 30 is diffused by the workpiece surface 35 as illustrated in FIG. 7A. Therefore, a bright area or highlight 62 and shadow areas 64 are formed within an image 60a inputted from the CCD camera 32 as shown in FIG. 7B. Thus, the histogram includes two peaks formed so as to correspond to the bright area 62 and the shadow areas 64 as shown in FIG. 8. Now, the average value of two luminances which form the two peaks, is regarded as a threshold value represented in a digitized level. less to say that the threshold value can by a method other than such an averaging luminances between the luminances of the used.
It is needalso be set method if two peaks are When the workpiece surface 35 is planar, the inspection light projected or emitted from the condensing optical system 30 is suitably introduced into the CCD camera 32 as illustrated in FIG. 9A. Thereafter, an image 60b (see FIG. 9B) inputted from the CCD camera 32 is represented in the form of a highlight or bright area 62 alone. Accordingly, the histogram includes only a peak formed so as to correspond to the bright area 62 as shown in FIG. 10. In this case, a threshold value a represented in a digitized level, is determined from the following equation 19 using the maximum luminance LMAX, which is expressed in the histogram, corresponding to the bright area 62, and the minimum luminance LMIN detected owing to the presence of points where a failure in the coating on the workpiece surface 35 is developed.
a = LMIN + (LMAX - LMIN)x A In this equation, A represents a suitable coefficient (O<A<1). For example, A may be set equal to 0.7 (i.e., A = 0. 7).
When, on the other hand, defective points are not developed on the workpiece surface 35 and only the peak of the bright area 62 appears, the maximum luminance LMAX is represented as the luminance of the peak and the minimum luminance LMIN is represented as 0.
In the present embodiment, when the workpiece surface 35 is curved upon setting of the digitized level, the value obtained by averaging the luminances corresponding to the respective peaks of the histogram, which are associated with the bright area 62 and the shadow areas 64 of the input image 60a, is set as the threshold value represented in the digitized level. It is therefore possible to prevent each of the shadow areas 64 from being misjudged as the bright area 62 upon digitization and to avoid the detection of the portion misjudged as a faulty or defective area upon subsequent image processing. Further, the threshold value is computed based on the maximum luminance LMAX and the minimum luminance LMIN Of the input image 60a when the workpiece surface 35 is flat. It is therefore possible to set a suitable threshold value. Thus, in the present embodiment, the threshold values can be automatically changed and established with respect to the input images 60a, 60b, thereby making it possible to avoid a problem that the inspection time increases.
The summary of the workpiece surface inspecting method will next be described with reference to FIG. 11.
An image on the workpiece surface 35 is first read by the CCD camera 32. The read image is then supplied to the image tint memory 42 of the image processing device 36 as tone image data. Data about a decision which has been made as to whether the workpiece surface 35 is plane, is inputted to the image processing device 36 from the robot controller 38. This decision can be made by judging whether or not the curvature of the workpiece surface 35 to be inspected is more than or equal to a preset curvature. The curvature of the workpiece surface 35 may be stored as teaching data.
When the workpiece surface 35 is planar, the tone image data is binary digitized by the image processor 44 and the digitized image data are stored in the digitized image memory 46.
21 Further, noise which has entered into the digitized image data and whose level is less than or equal to a predetermined level, is removed by an electrical process.
Then, isolated spots or points (isolated black portions on an image) are detected from the image data obtained in the above-described manner. The number of the isolated points is counted and the area of each isolated point is measured. It is thereafter determined, based on the results of counting and measurement, whether or not the respective isolated points are points which are inferior in coating due to dust or the like. The result of determination is outputted to the monitor TV 50.
When the workpiece surface 35 is curved, input image data is binary digitized in the same manner as when the workpiece surface 35 is planar. Then, noise whose level is less than or equal to a given level, is removed by an electrical process. In addition to such a noise removing method, the present invention discloses a window setting method or a method of reducing portions to be processed. These methods will hereinafter be described in detail.
The window setting method (i.e., the routine executed in either Step S24 or Step S26 selected in Step S10 shown in FIG. 5) used when the workpiece surface 35 is curved, will first be described below.
22 The routine of Step S24 will hereinafter be described in detail with reference to FIGS. 12 through 14.
When the workpiece surface 35 is curved, every block inspected by the CCD camera 32 has different effective inspection regions. When the surface of a vehicle hood or bonnet is inspected, for example, as illustrated in FIG. 12, bright and shadow areas of read images I through IV are different in shape from each other depending on spots on the bonnet to be inspected. Accordingly, a plurality of windows shown in FIG. 13 is set in advance with respect to such various images in order to delete image processing spots of the shadow areas, i.e. areas other than the effective inspection regions. As shown in the flowchart of FIG. 14, the robot controller 38 first inputs
a predetermined window No. to the image processing device 36 in accordance with data about a point where an image on a workpiece is being read (Step S40). Then, an image inputted to the image processing device 36 from the CCD camera 32 is stored in the image tint memory 42 as tone image data. Thereafter, the image processor 44 is activated to set the window selected in Step S40 so as to correspond to the tone image data (Step S42). When the image I shown in FIG. 12 is taken in, for example, the window No. 1 depicted in FIG. 13 is selected and set up.
23 When, on the other hand, the image IV is taken in, the window No. 6 is selected -and set up. By setting up the windows in this way, it is unnecessary to perform the image processing in Step 34 and that subsequent to Step S34 on the majority of the shadow areas (portions indicated by oblique lines) of the taken-in images I through IV.
Then, image data in the windows set to the images I through IV are binary digitized in accordance with the digitized level which has been set in Step S8 of FIG. 5 (Step S44). Further, the digitized image data are inputted to the digitized image memory 46 (Step S46). After the respective windows have been set to the images I through IV in this way, only the images within the windows are digitized and the image processing in Step 34 and that subsequent to Step S34 of FIG. 5, is executed.
Thus, the time required to process the images can be greatly reduced because the respective windows are set to the images I through IV, and the shadow areas (i.e. areas other than the effective inspection regions) are deleted from the image processing.
Then, the routine to be executed in Step S26 will be described below with reference to FIGS. 15 through 19.
An image inputted to the image processing device 36 from the CCD camera 32 is first digitized in accord- 24 ance with the digitized level which has been set in Step S8 of FIG. 5. Then, the digitized image whose bright and shadow spots or areas are represented as 1 (white) and 0 (black) respectively, is inputted to the digitized image memory 46 as image data (Step S50).
In the image processor 44, an x direction and a y direction are set as illustrated in FIG. 16A. In this condition, the number of pixels in the x direction, which correspond to bright areas of digitized image data and the number of pixels in the y direction, which correspond to bright areas of the image data, are added together so as to produce projected values, thereby creating a histogram based on the soproduced projected values (Step S52). A value Xdif corresponding to fmaximum projected value minimum projected value) in the x direction of the histogram is compared with a value Ydif corresponding to (maximum projected value - minimum projected value) in the y direction thereof (Stop S54). This comparison is effected to determine whether an actually-set window extends in the x or y direction. When the digitized image data shown in FIG. 16A is used, for example, Xdif is less than Ydif (i.e., Xdif'Ydif). Accordingly, the window to be set extends in the x direction.
Even if digitized image data are represented in diagonal form shown in FIG. 17A, there is often a situation in which Xdif is less than Ydif (i.e. , Xdif'ydif). At this time, the shape of a window is set as shown in FIG. 17B. As this is identical to the case where the window is not set, it is therefore necessary to bypass the process for setting the window (as described in Steps S58 through S62, later).
It is determined whether or not the highlight or bright area of the digitized image is represented in diagonal form (Step S56).
If the answer is determined to be "No" in Step S56, then a window setting region is computed. The region is set so as to extend to the limit of an image region along the x direction and to the end of the bright area along the y direction (Step S58).
Further, a window region set to the boundary between the bright area and the shadow area, is increased several pixels toward the shadow area side (Step S60). The increase in the window region is made to avoid a situation in which information about dust or the like which falls within the boundary between the bright area and the shadow area, cannot be detected.
After the window region has been set in this way, a window shown in FIG. 16B is set so as to correspond to the image (Step S62).
If, on the other hand, the answer is determined to be "Yes" in Step S56, then the entire image should be regarded as being an object for image processing.
26 Therefore, the setting of a window is not effected.
Similarly, if Xdif'Ydif in Step S54 as shown in FIG. 18A, then the routine similar to that executed in each of Steps S56, S58 is executed in each of Steps S64, S66. Further, the routine of each of Steps S60, S62 is executed, after which a window shown in FIG. 18B is set so as to correspond to an image.
Note that "diagonal" herein means joining two nonadjacent corners of a rectangle, so that in FIG. 19A, the bright area is not represented in diagonal form and therefore, in accordance with the present embodiment, the image of FIG. 19A will be associated with the window of FIG. 19B, and accordingly, the shadow area to be image processed is reduced.
In the present embodiment as described above, the windows corresponding to the respective input images are set, except where the bright area is represented in diagonal form. It is therefore unnecessary to perform a dilation and erosion process to be described later on the shadow area. When the bright area is represented in diagonal form, the window setting process (Steps S58, S66 and S60) is not executed, thereby making it possible to reduce the time required to process each image.
The method (i.e., the routine executed in either Step S28 or Step S32 selected in Step S10 shown in FIG. 3) of combining the images obtained when the 27 workpiece surface 35 is curved, will next be described in detail.
The image combining routine of Step S28 will be described below with reference to FIGS. 3 through 5 and FIG. 20.
In this case, all the storage regions of binary memories No. 0 through No. 2 which constitute the digitized image memory 46, are initially set to 0 (black).
Then, an image at a first block is inputted to the image processing device 36 from the CCD camera 32. The operation of the arm 26 of the robot 22 and the position of the CCD camera 32 are controlled, upon teaching the robot 22, by the robot controller 38, thereby positionally adjusting a bright spot or area so as to be brought to an upper portion of a frame of the CCD camera 32. The input image is first stored in the image tint memory 42 as tone image data. Thereafter, the image is by- level digitized based on the digitized level which has been set in Step S8 shown in the flowchart of FIG. 5. The digitized image whose bright and shadow spots or areas are represented as 1 (white) and 0 (black) respectively, is stored in the binary memory No. 0 of the digitized image memory 46.
Further, the images stored in the binary memories No. 2 and No. 0 of the digitized image memory 46 are ORed or OR-operated for each pixel by the image 28 processor 44 to produce a new composite image, which is in turn stored in the binary memory No. I of the digitized image memory 46.
Then, the CCD camera 32 is displaced by the robot controller 38. Thereafter, an image at a adjacent second block is brought into the image processing device 36 from the CCD camera 32. In this case, the bright area of the image is positionally brought to the center of the frame of the CCD camera 32 taught by the robot controller 38 to so perform. That is, the bright area of the image is set so as not to be superimposed on the bright area of the image at the first block. Then, the image is digitized so as to represent the bright and shadow areas thereof as 1 (white) and 0 (black) respectively. Thereafter, the digitized image is re- inputted to the binary memory No. 0.
Then, the composite image stored in the binary memory No. 1 and a new image inputted to the binary memory No. 0 are OR-operated for each pixel by the image processor 44 so as to produce a combined image. In this case, the bright areas of the images at the first and second blocks are not superimposed on one another by the robot controller 38 taught to so perform. Thus, as is apparent from FIG. 20 schematically illustrating the combination of the images, the bright areas of the respective images are stored as data in the binary memory No. 2 without being superimposed on 29 one another.
Similarly, the robot 22 is activated to displace the arm 26 to thereby enable an image at an adjacent third block to be inputted to the binary memory No. 0 from the CCD camera 32. Then, the composite image which has been stored in the binary memory No. 2, is combined with a new image inputted to the binary memory No. 0 to produce a combined image. The image thus combined is stored in the binary memory No. 1. Even in this case, the bright areas of the images of the first, second and third blocks are stored as data as shown in FIG. 20 without being superimposed on one another.
Thus, the images to be inspected on the curved workpiece surface, are combined together in a continuous state of the first through third blocks into one image. Further, tha image thus combined is subjected to the image processing as a single image. It is therefore possible to reduce the area of unnecessary shadow area to be subjected to the image processing and to greatly improve image processing speed as compared with the case where the three images are separately subjected to the image processing.
Incidentally, while the three images are combined into one in the present embodiment, it is, however, needless to say that the number of images to be combined into one image can be changed according to the widths of the bright areas.
Next, the image combining routine of Step S32 will be described below with reference to FIGS. 3 through 5, and a flowchart and a view for describing the combination of images, both of which are shown in FIG. 21.
First, all the storage regions of the binary memories No. 0 through No. 2 which constitute the digitized image memory 46, are initially set to 0 (black)(Step S2 in FIG. 5). Then, an image at a first block is inputted to the image processing device 36 from the CCD camera 32. The input image is first stored in the image tint memory 42 as tone image data. Thereafter, a digitized level is set in Step S8 of the flowchart shown in FIG. 5. Further, the image is binary digitized based on the digitized level. The digitized image whose bright and shadow areas are represented as I (white) and 0 (black) respectively, is stored in the binary memory No. 1 (Step S70).
Then, an x direction and a y direction are set as illustrated in FIG. 21 with respect to the binary memory No. 1. In this condition, the number of pixels in the x direction, which correspond to a bright spot or area of the digitized image and the number of pixels in the y direction, which correspond to the bright area of the image, are added together so as to produce projected values, thereby creating a histogram 31 based on the projected values. Then, a window is set based on the histogram so as to correspond to a portion equivalent to the bright area. Further, the image is divided into the bright and shadow areas by using the window again, followed by being re-inputted to the binary memory No. 2 (Step S72). It is next determined whether or not the remaining region of the binary memory No. 0 inputted with the composite image, is greater than a region delineated by a widthz\y which extends in the y direction of the window (Step S74). If the answer is determined to be "Yes" in Step S74, then the image of the binary memory No. 0 and the image of the binary memory No. 2 are OR-operated. In this case, the image of the binary memory No. 2 is combined with the image of the binary memory No. 1 in a state in which the image of the binary memory No. 2 has been shifted to the top as seen in the drawing. The image thus combined is re-inputted to the binary memory No. 1 (Step S76). Similarly, it is determined (in Step S78) whether or not there is enough remaining region in binary memory No. 1 for enabling the images to be combined again. The remaining region is compared with a predetermined region. If the answer is determined to be "Yes" in Step S78, then the image combining process to be executed in Step S70 and Steps subsequent to Step S70, is repeated. When a second image combining process and image combining processes 32 z subsequent to the second image combining process, are executed, an image corresponding to the window reformed in Step S72 is not superimposed on the image corresponding to the window formed prior to the present window in Step S76 for executing the image combining process. Further, the image produced in Step S72 is shifted so as not to be spaced from the image corresponding to the window formed prior to the present window, and combined with the latter image. Thus, an image subjected to the combination four times, for example, is continuously formed based on the first to fourth image combining processes as shown below in FIG. 21.
When, on the other hand, it is determined in Step S74 that the width y of the window is more than that of the remaining region of the binary memory No. 1 or when it is determined in Step S78 that the width of the residual region of the newly combined image makes it impossible to further effect an image combining process, the shifting image combining process is completed. The same effect as that obtained when the image combining process is executed in Step S28 can be brought about by effecting the above process in this way. Further, bright areas of a number of images can be combined into one by displacing the position of the window upon combination of the images. Therefore, the memory capacity can be effectively used and the image 33 processing speed can be improved.
Thus, when the workpiece surface is curved, the image subjected to the image processing is reduced and the dilation and erosion process is thereafter executed.
An image 60c, which has been subjected to the digitization and the noise elimination, is represented as shown in FIG. 22. That is, the state of coating of the workpiece surface has an influence on an image to be detected because the workpiece surface 35 is curved, and the boundary between a bright area 62 and a shadow area 64 is indefinite. Further, a number of isolated points or spots 66 are included in the bright area 62.
The digitized image data shown in FIG. 22 is first processed for erosion to thereby produce an image 60d excluding unnecessary isolated points 66, each of which is not a faulty or defective point and is represented at a level below a given level, as shown in FIG. 23. Further, the image 60d is processed for dilation. As a result, the boundary between the bright area 62 and the shadow area 64 is rendered definite, thereby enabling the image 60d to be represented as an image 60e shown in FIG. 24. Further, a point or spot 68 such as dust or the like inferior in coating, which has been eroded by erosion, is dilated and hence rendered definite.
34 1 1 As described above, the isolated points such as the faulty point 68, etc., are detected from the image which has been processed for dilation and erosion in the same manner as when the workpiece surface 35 is planar. Then, the number of the isolated points is counted and the area of each isolated point is measured. It is also determined, based on the results of both the counting and the measurement, whether or not the respective isolated points are points such as dust, etc., which are inferior in coating. The result of determination is then outputted to the monitor TV 50.
According to the workpiece surface inspecting method of the present embodiment, as has been described above, when a workpiece surface is curved, defective points can be reliably detected by processing image data for dilation and erosion. Further, isolated points are no longer misdetected as the defective points. When, on the other hand, the workpiece surface is planar, the time required to process an image can be reduced by omitting a dilation and erosion process.
Having now fully described the invention, it will be apparent to those skilled in the art that many changes and modifications can be made without depart ing from the scope of the invention as set forth herein.

Claims (8)

CLAIMS:
1. A method of inspecting a surface of a workpiece by using a surface inspecting apparatus provided with projecting means for illuminating the workpiece surface to be inspected with detection light, detecting means for receiving and detecting the light reflected from the workpiece surface, and a condensing optical system for causing said reflected light to converge on a light-receiving surface of said detecting means, comprising the following steps of: successively displacing said surface inspecting apparatus along the workpiece surface, receiving and detecting said reflected light by said detecting means and holding image data obtained from said detecting means; binary digitizing said image data; determining whether or not the workpiece surface to be inspected is planar; and processing the digitized image data for dilation and erosion when the workpiece surface to be inspected is non-planar.
2. A method of inspecting a surface of a workpiece by using a surface inspecting apparatus provided with projecting means for irradiating the workpiece surface to be inspected with detection light, detect- 36 ing means for receiving and detecting the light reflected from the workpiece surface, and a condensing optical system for causing said reflected light to converge on a light-receiving surface of said detecting means, comprising the following steps of: selecting a preset window for extracting, based on the shape of the workpiece surface to be inspected, an effective image to be inspected from an image obtained by said detecting means; and performing image processing on an image included in an effective inspection region, said image having been extracted by using said selected window.
3. A method of inspecting a surface of a workpiece by using a surface inspecting apparatus provided with projecting means for illuminating the workpiece surface to be inspected with detection light, detecting means for receiving and detecting the light reflected from the workpiece surface, and a condensing optical system for causing said reflected light to converge on a light-receiving surface of said detecting means, comprising the following steps of: binary digitizing an image obtained by said detecting means; producing a histogram graphically representing a relationship between bright areas and shadow areas from said digitized image; setting a window for extracting an effective image to be inspected, based on said histogram; extracting an effective image to be inspected from the digitized image by using said set window; and performing image processing on said extracted image.
4. A method of inspecting a surface of a workpiece by using a surface inspecting apparatus provided with projecting means for illuminating the workpiece surface to be inspected with detection light, detecting means for receiving and detecting the light reflected from the workpiece surface, and a condensing optical system for causing said reflected light to converge on a light-receiving surface of said detecting means, comprising the following steps: a first step for successively displacing said surface inspecting apparatus along the workpiece surface, receiving and detecting said reflected light by said detecting means and by-level digitizing a plurality of images obtained by said detecting means; a second step for extracting image data corresponding to an effective inspection surface of the workpiece irradiated with said detection light from said plurality of digitized image data and for holding said extracted image data in a non-superimposed state; and 38 a third step for subjecting said held image data to given image processing.
5. A method according to claim 4, further comprising a fourth step for teaching said surface inspecting apparatus in advance so as to avoid the superimposition of said plurality of image data on one another.
to define 2
6. A method according to claim 4 or clairr 15, wherein said second step includes a step for setting a window used region corresponding to the effective inspection surface, based on the extracted digitized image data and a step for holding image data corresponding to the effective inspection surface in a nonsuperimposed state, said image data having been extracted by said window.
7. A method of inspecting a surface of a workpiece by using a surface inspecting apparatus provided with projecting means for illuminating the workpiece surface to be inspected with detection light, detecting means for receiving and detecting the light reflected from the workpiece surface, and a condensing optical system for causing said reflected light to converge on a light-receiving surface of said detecting means, comprising the following steps of:
39 1 successively displacing said surface inspecting apparatus along the workpiece surface, receiving and detecting said reflected light by said detecting means and holding images obtained by said detecting means as tone image data every pixels; creating a histogram graphically representing the number of pixels corresponding to respective tone levels based on said tone image data; setting a threshold value represented in a digitized level between the maximum value of a tone level in said histogram as seen on the bright area side and the maximum value of a tone level in said histogram as seen on the shadow area side; converting said tone image data into digitized image data based on said threshold value; and subjecting said digitized image data to image processing.
8. A method of inspecting a surface of a workpiece substantially as hereinbefore described with reference to Figures 3 to 24.
GB9225600A 1991-12-13 1992-12-08 Method of inspecting the surface of a workpiece Expired - Fee Related GB2262339B (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
JP3329384A JPH05164700A (en) 1991-12-13 1991-12-13 Method for inspecting work surface
JP3329383A JP2941107B2 (en) 1991-12-13 1991-12-13 Work surface inspection method
JP3333404A JPH05164703A (en) 1991-12-17 1991-12-17 Inspecting method for surface of workpiece
JP3337263A JP2904983B2 (en) 1991-12-19 1991-12-19 Work surface inspection method

Publications (3)

Publication Number Publication Date
GB9225600D0 GB9225600D0 (en) 1993-01-27
GB2262339A true GB2262339A (en) 1993-06-16
GB2262339B GB2262339B (en) 1995-09-06

Family

ID=27480434

Family Applications (1)

Application Number Title Priority Date Filing Date
GB9225600A Expired - Fee Related GB2262339B (en) 1991-12-13 1992-12-08 Method of inspecting the surface of a workpiece

Country Status (2)

Country Link
US (1) US5379347A (en)
GB (1) GB2262339B (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2270227A (en) * 1992-08-28 1994-03-02 Surface Inspection Ltd Linescan visual monitoring system
FR2707008A1 (en) * 1993-06-21 1994-12-30 Peugeot Method for inspecting a weld bead, in particular on a motor-vehicle suspension sphere and device for implementing this method
US5566244A (en) * 1993-11-22 1996-10-15 Honda Giken Kogyo Kabushiki Kaisha Method of inspecting a workpiece surface including a picturing system with a shortened focal plane
EP1832868A1 (en) * 2006-03-10 2007-09-12 Omron Corporation Device for and method of inspecting surface condition
EP2770322A1 (en) * 2013-02-26 2014-08-27 C.R.F. Società Consortile per Azioni Method and system for detecting defects in painting of components, in particular of motor-vehicle bodies
EP2799847A1 (en) * 2013-05-03 2014-11-05 Compar AG Method for optically testing the surfaces of objects
IT201900021645A1 (en) * 2019-11-19 2021-05-19 Sacmi APPARATUS FOR THE OPTICAL INSPECTION OF HEALTHCARE

Families Citing this family (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5517027A (en) * 1993-06-08 1996-05-14 Mitsubishi Denki Kabushiki Kaisha Method for detecting and examining slightly irregular surface states, scanning probe microscope therefor, and method for fabricating a semiconductor device or a liquid crystal display device using these
JP3258821B2 (en) * 1994-06-02 2002-02-18 三菱電機株式会社 Method for positioning and analyzing minute foreign matter, analyzer used for the method, and method for manufacturing semiconductor element or liquid crystal display element using the same
US5878153A (en) * 1994-06-24 1999-03-02 Ford Global Technologies, Inc. Method for monitoring coating adhesion propensity based on surface morphology
US5732147A (en) * 1995-06-07 1998-03-24 Agri-Tech, Inc. Defective object inspection and separation system using image analysis and curvature transformation
EP0880023A1 (en) * 1997-05-23 1998-11-25 Siemag Transplan Gmbh Method and device for the automatic detection of surface faults during the continuous mechanical removal of material from casted products
US6294327B1 (en) * 1997-09-08 2001-09-25 Affymetrix, Inc. Apparatus and method for detecting samples labeled with material having strong light scattering properties, using reflection mode light and diffuse scattering
EP1173749B1 (en) * 1999-04-19 2005-09-28 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Image editing for preparing a texture analysis
US6266138B1 (en) * 1999-10-12 2001-07-24 Perceptron, Inc. System and method for detecting defects in a surface of a workpiece
JP4743805B2 (en) * 2000-04-06 2011-08-10 ルネサスエレクトロニクス株式会社 Appearance inspection method and apparatus
DE10036741A1 (en) 2000-07-27 2002-02-07 Duerr Systems Gmbh Process and control system for checking the coating quality of workpieces
JP2003148914A (en) * 2001-11-08 2003-05-21 Fanuc Ltd Position detector and taking-out device using position detection
JP2004048379A (en) * 2002-07-11 2004-02-12 Murata Mach Ltd Image processing apparatus
DE102004007829B4 (en) * 2004-02-18 2007-04-05 Isra Vision Systems Ag Method for determining areas to be inspected
NZ533533A (en) * 2004-06-11 2006-11-30 Avalon Engineering Ltd Deployment method and apparatus for moving stacked articles in varying orientations
US7253908B2 (en) * 2004-07-22 2007-08-07 The Boeing Company Non-destructive inspection using laser profiling and associated method
JP4652024B2 (en) * 2004-11-22 2011-03-16 富士通株式会社 Surface inspection method and apparatus
JP4985394B2 (en) * 2005-03-15 2012-07-25 オムロン株式会社 Image processing apparatus and method, program, and recording medium
JP6750841B2 (en) * 2015-06-26 2020-09-02 キヤノン株式会社 Inspection method, inspection device, processing device, program and recording medium
JP6915288B2 (en) * 2017-02-08 2021-08-04 オムロン株式会社 Image processing system, image processing device, circuit reconstruction method in FPGA (Field Programmable Gate Array), and circuit reconstruction program in FPGA
CN107328789B (en) * 2017-06-23 2023-09-26 宁波韵升智能技术有限公司 Arc surface and curved surface defect detection system
CN108876784B (en) * 2018-06-27 2021-07-30 清华大学 Image processing method and device for removing connecting part of planar workpiece

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB1549706A (en) * 1975-07-11 1979-08-08 Armines Equipment for analysing at least one heterogeneous two-or three-dimensional medium
GB2032618A (en) * 1978-09-18 1980-05-08 Eastman Kodak Co Inspection of elongate articles for surface irregularities
GB2221032A (en) * 1988-06-29 1990-01-24 Gd Spa Method of inspecting the ends of stacked cigarettes
EP0371650A1 (en) * 1988-11-28 1990-06-06 Allegheny Ludlum Corporation Slab surface contour monitor
US4980923A (en) * 1986-10-22 1990-12-25 Seiko Instruments Inc. Dilation/erosion conversion circuit
GB2239089A (en) * 1987-10-08 1991-06-19 Elor Optronics Ltd Small arms ammunition inspection system

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE3215673A1 (en) * 1981-04-29 1982-11-25 Mitsubishi Denki K.K. DESIGN SCANNER
JPS5821146A (en) * 1981-07-30 1983-02-07 Kirin Brewery Co Ltd Method and device for inspection of defect
US4542527A (en) * 1982-05-17 1985-09-17 Tokyo Shibaura Denki Kabushiki Kaisha Image processing device for continuously extracting features of small regions of an image
CA1250050A (en) * 1985-04-17 1989-02-14 Miyahiko Orita Image processing apparatus
IL79098A0 (en) * 1985-07-15 1986-09-30 Visionetics Corp Fine flaw detector for printed circuit boards
US4799268A (en) * 1985-11-12 1989-01-17 Usm Corporation Lead sense system for component insertion machine
ATE109278T1 (en) * 1986-10-03 1994-08-15 Omron Tateisi Electronics Co DEVICE FOR INVESTIGATION OF AN ELECTRONIC DEVICE IN FIXED ASSEMBLY.
US4868404A (en) * 1987-04-23 1989-09-19 Hajime Industries, Ltd. Surface inspection apparatus using a mask system to monitor uneven surfaces
JPH07119709B2 (en) * 1988-01-13 1995-12-20 トヨタ自動車株式会社 Surface defect inspection device
JPH0672770B2 (en) * 1988-02-01 1994-09-14 豊田工機株式会社 Robot object recognition device
JPH01280977A (en) * 1988-03-31 1989-11-13 Toshiba Corp Method and apparatus for displaying slave picture of television system
JPH0244202A (en) * 1988-08-05 1990-02-14 Bridgestone Corp Apparatus for detecting end position of object
US4958223A (en) * 1988-09-16 1990-09-18 Owens-Brockway Glass Container Inc. Inspection of container finish
US4975972A (en) * 1988-10-18 1990-12-04 At&T Bell Laboratories Method and apparatus for surface inspection
GB9008632D0 (en) * 1990-04-17 1990-06-13 Leicester Polytechnic Inspecting garments
US5237404A (en) * 1990-06-28 1993-08-17 Mazda Motor Corporation Inspection apparatus with improved detection of surface defects over large and curved surfaces

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB1549706A (en) * 1975-07-11 1979-08-08 Armines Equipment for analysing at least one heterogeneous two-or three-dimensional medium
GB2032618A (en) * 1978-09-18 1980-05-08 Eastman Kodak Co Inspection of elongate articles for surface irregularities
US4980923A (en) * 1986-10-22 1990-12-25 Seiko Instruments Inc. Dilation/erosion conversion circuit
GB2239089A (en) * 1987-10-08 1991-06-19 Elor Optronics Ltd Small arms ammunition inspection system
GB2221032A (en) * 1988-06-29 1990-01-24 Gd Spa Method of inspecting the ends of stacked cigarettes
EP0371650A1 (en) * 1988-11-28 1990-06-06 Allegheny Ludlum Corporation Slab surface contour monitor

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2270227A (en) * 1992-08-28 1994-03-02 Surface Inspection Ltd Linescan visual monitoring system
FR2707008A1 (en) * 1993-06-21 1994-12-30 Peugeot Method for inspecting a weld bead, in particular on a motor-vehicle suspension sphere and device for implementing this method
US5566244A (en) * 1993-11-22 1996-10-15 Honda Giken Kogyo Kabushiki Kaisha Method of inspecting a workpiece surface including a picturing system with a shortened focal plane
EP1832868A1 (en) * 2006-03-10 2007-09-12 Omron Corporation Device for and method of inspecting surface condition
US7782451B2 (en) 2006-03-10 2010-08-24 Omron Corporation Device for and method of inspecting surface condition having different curvatures
EP2770322A1 (en) * 2013-02-26 2014-08-27 C.R.F. Società Consortile per Azioni Method and system for detecting defects in painting of components, in particular of motor-vehicle bodies
EP2799847A1 (en) * 2013-05-03 2014-11-05 Compar AG Method for optically testing the surfaces of objects
IT201900021645A1 (en) * 2019-11-19 2021-05-19 Sacmi APPARATUS FOR THE OPTICAL INSPECTION OF HEALTHCARE
WO2021099949A1 (en) * 2019-11-19 2021-05-27 Sacmi Cooperativa Meccanici Imola Societa' Cooperativa Apparatus for optical inspection of sanitaryware

Also Published As

Publication number Publication date
US5379347A (en) 1995-01-03
GB9225600D0 (en) 1993-01-27
GB2262339B (en) 1995-09-06

Similar Documents

Publication Publication Date Title
US5379347A (en) Method of inspecting the surface of a workpiece
JP4172761B2 (en) Defect inspection apparatus, defect inspection method and program
US7433508B2 (en) Pattern inspection method and its apparatus
US6674890B2 (en) Defect inspection method and apparatus therefor
US7274813B2 (en) Defect inspection method and apparatus
US7336815B2 (en) Image defect inspection method, image defect inspection apparatus, and appearance inspection apparatus
JP2006138708A (en) Image flaw inspection method, image flaw inspecting device and visual inspection device
US4779002A (en) Method and apparatus for measuring optical cutting beam with parallel beam referencing
JPH05164703A (en) Inspecting method for surface of workpiece
JP2003065969A (en) Apparatus and method for inspecting pattern
JP3374818B2 (en) Defect detection apparatus and method
JP2002303588A (en) Pattern defect inspection device
JP3055322B2 (en) Circular container inner surface inspection device
JP3025562B2 (en) Surface condition inspection method using bright and dark illumination
JP3055323B2 (en) Circular container inner surface inspection device
JPH06213820A (en) Foreign-matter discriminating method
JP2874402B2 (en) Circular container inner surface inspection device
JP3190173B2 (en) Semiconductor device appearance inspection method and appearance inspection device
JPS61126455A (en) Image processing method
JPH04282442A (en) Pattern inspection apparatus
JPH05164700A (en) Method for inspecting work surface
JP3545864B2 (en) Pattern defect inspection equipment
JP2941107B2 (en) Work surface inspection method
JPH07159141A (en) Inspection of surface of workpiece
JPH03215706A (en) Pattern inspecting device

Legal Events

Date Code Title Description
PCNP Patent ceased through non-payment of renewal fee

Effective date: 20011208