WO2019177539A1 - Method for visual inspection and apparatus thereof - Google Patents

Method for visual inspection and apparatus thereof Download PDF

Info

Publication number
WO2019177539A1
WO2019177539A1 PCT/SG2019/050138 SG2019050138W WO2019177539A1 WO 2019177539 A1 WO2019177539 A1 WO 2019177539A1 SG 2019050138 W SG2019050138 W SG 2019050138W WO 2019177539 A1 WO2019177539 A1 WO 2019177539A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
defect
capturing device
image capturing
pattern
Prior art date
Application number
PCT/SG2019/050138
Other languages
French (fr)
Inventor
Jierong CHENG
Jia Du
Wei Xiong
Wenyu Chen
Joo Hwee Lim
Original Assignee
Agency For Science, Technology And Research
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Agency For Science, Technology And Research filed Critical Agency For Science, Technology And Research
Priority to SG11202008944XA priority Critical patent/SG11202008944XA/en
Publication of WO2019177539A1 publication Critical patent/WO2019177539A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0004Industrial image inspection
    • G06T7/001Industrial image inspection using an image reference approach
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/50Extraction of image or video features by performing operations within image blocks; by using histograms, e.g. histogram of oriented gradients [HoG]; by summing image-intensity values; Projection analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30108Industrial image inspection
    • G06T2207/30164Workpiece; Machine component
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/06Recognition of objects for industrial automation

Definitions

  • the invention relates to a method for visual inspection and an apparatus thereof, in particular, the visual inspection involves detecting a defect on an object.
  • Computer vision is a sub field of artificial intelligence that investigate how to make computer algorithms perceive and understand the world of a human through images.
  • light is detected by cameras and/or sensors, while high-level of image understanding is processed by using computer algorithms.
  • computer vision plays an important role. For instance, computer vision can be used for range sensing, industrial inspection of manufactured parts and/or parts that are currently in use and susceptible to wear and tear, as well as object recognition.
  • Visual inspection of surfaces for defects can cover both geometry (dimension) inspection and surface (texture) inspection of the surfaces.
  • Dimension inspection is based on geometric shapes and typically captured using a positioning device holding a geometry sensor (a depth sensor, or a 3D scanner and etc.), while surface (texture) inspection is based on images.
  • Sensor planning i.e. identification of sensor locations to capture different parts of an object with required specifications is important for acquiring desired geometric shapes for dimension inspection.
  • surface inspection relies on the imaging quality of camera(s) used. Different views of a target region may need to be acquired for analysis to identify the existence of a defect.
  • viewpoints for all these systems are predefined before the sensing, e.g. with active sensor planning.
  • active sensor planning e.g., in surface inspection of objects of the same type, manual selection is required to define viewpoints for image capturing, while in dimension inspection systems, CAD-based active sensing is adopted to derive viewpoints for capturing the 3D geometry.
  • Multi-view based inspection systems may acquire inspection data by combining both shapes and images from different viewpoints.
  • computer vision is particularly valuable in 3D construction and/or reconstruction because computer vision can be used to obtain a three-dimensional position of an object point from its projective points on the image plane.
  • a 3D construction and/or reconstruction is necessary in applications where accuracy of the 3D shape represents a fundamental factor to the performance of a computer vision device.
  • high-end 3D scanners are required to achieve a good 3D reconstruction of a target object under inspection.
  • Figure 1 A shows how a crack is observed in a hash-outlined box and how a crack is occluded in a black-outlined box in one captured photograph.
  • Figure 1 B shows how a crack is occluded in the hash-outlined box of Figure 1 A in another captured photograph.
  • Figure 1 C shows how a crack is observed in the black-outlined box of Figure 1 A in another captured photograph.
  • Figure 2 shows a flow chart of how a method for performing single-view mobile inspection of defects can be provided according to one example of the present disclosure.
  • Figure 3 is a geometry illustration for detection of a hole crack according to one example of the present disclosure.
  • Figure 4A shows a test image comprising a crack region that is marked by a visual inspector.
  • Figure 4B is an exploded view of the crack region in Figure 4A.
  • Figure 4C shows a plurality of detected cracks in the test image according to one example of the present disclosure.
  • Figure 5 shows a flow chart of how classification-based single-view inspection is learnt from training images according to one example of the present disclosure.
  • Figure 6A shows how a plurality of image patches are obtained from a captured image according to an example of the present disclosure.
  • Figure 6B shows how erosion/corrosion regions can be indicated as on a captured image according to one example of the present disclosure.
  • Figure 7 shows a flow chart for single-view inspection trained with a 2D reference image according to one example of the present disclosure.
  • Figure 8 shows a flow chart for single-view inspection trained with a 3D reference model according to one example of the present disclosure.
  • Figure 9A shows a sensor planning system for computer vision according to one example of the present disclosure.
  • Figure 9B shows a comparative example of an existing model of sensor planning system for computer vision.
  • Figure 10 shows a work flow of a proposed Snap-to-inspect three Dimensional system (Snap2lnspect3D) according to one example of the present disclosure.
  • Figure 1 1 shows a hardware set up for capturing an image for visual inspection with active 3D measurement according to one example of the present disclosure.
  • Figure 12 shows how system calibration can be formulated according to one example of the present disclosure.
  • Figure 13A shows an example of a coded pattern according to one example of the present disclosure.
  • Figure 13B shows how a portion of the coded pattern in Figure 13A can be decoded.
  • Figure 13C shows how a cylinder is reconstructed after decoding the coded pattern in Figure 13A.
  • Figure 13D show an example of a checkerboard pattern that can be used for system calibration according to one example of the present disclosure.
  • Figure 13E show an example of asymmetrical circle patterns that can be used for system calibration according to one example of the present disclosure.
  • Figure 14 shows how a projection pattern can be generated for defect measurement according to one example of the present disclosure.
  • Figure 15 is a screenshot of system calibration with asymmetrical circle patterns according to one example of the present disclosure.
  • Figure 16 is an example of a system calibration result after system calibration in Figure 15.
  • Figure 17 is an example of height measurement of a rectangular object on a surface using Snap2lnspect3D.
  • Figure 18 is a flow chart on how Snap2lnspect3D operates according to one example of the present disclosure.
  • Figure 19 is a block diagram illustrating a system architecture according to an example of the present disclosure.
  • Figures 1 A to 1 C are images of a blade surface taken from different views, and these figures illustrate how occlusion and/or shading can hide defect features.
  • the crack in hashed-outlined box 101 can be observed in Figure 1 A, but hardly seen in Figure 1 B, and the crack in black-outlined box 102 can be clearly observed in Figure 1 C, but is very weak in Figure 1 A. Flence, it is important to use an appropriate view for optimal defect detection.
  • Mobile (portable or robotic) imaging devices enable flexible observation settings (angles, distances, etc.). A quick snap or image capture of the object surface without intentional planning using a mobile device is likely to result in a sub-optimal observation of defects.
  • Flandheld devices such as a smart phone or an imaging device that can be held by a user for capturing images
  • the picture(s) taken by a handheld device may not be at an optimal viewing angle and distance (or optimal observation settings) to distinguish an abnormity from the background or other content on a target object.
  • This technique imposes challenges to the demand of excellent feature representation and excellent detection algorithms which can detect abnormities under various imaging and lighting conditions.
  • a further challenge is resolving the scale of the target object where visual inspection tasks require measurements in absolute scale.
  • a solution for mobile inspection of defects assisted by a variety (or different types) of visual knowledge is provided to address the challenges described above.
  • Single view inspection used herein refers to inspection using only a single view for capturing three dimensional (3D) geometry and/or two dimensional (2D) images to enable dimension inspection and surface texture inspection. This can be referred as“one-shot scan” in the present disclosure.
  • Portable used herein refers to a lighter and smaller version (than usual) of an object that allows it to be easily carried or moved or transported around from one place to another place by a user.
  • Engine blade used herein may refer to a turbine blade of a gas turbine or a steam turbine, or may refer to fan blades or propellers used in an automotive vehicle (such as a jet), industrial condensers, industrial fans and the like.
  • a method and a device for single-view mobile inspection of defects assisted by different types of visual knowledge In one aspect of the exemplary technique, visual knowledge about the defect characteristics, such as training images or reference image, is acquired as part of input of the defect detection algorithms. In another aspect of the exemplary technique, defects are detected from individual single-view images captured by an image capturing device, preferably hand held by a user conducting the inspection, and concurrently, a confidence score is calculated, and returned as feedback to the user. In another aspect of the exemplary technique, better inspection results are incrementally achieved when imaging parameters are tuned, preferably by the user.
  • a method, (vision) system and/or apparatus for performing single-view mobile inspection of defects that is guided by different types of visual knowledge.
  • These method, system and/or apparatus are termed in the present disclosure as“snap- to-inspect”, “snap2inspect2D” or “snap2inspect”.
  • An extension of these method, system and/or apparatus is termed as“snap2inspect3D”.
  • a flow chart illustrating the operation principles of the snap- to-inspect example is illustrated in Figure 2.
  • an image capturing unit e.g. camera, sensor and the like
  • a mobile device such as smartphone
  • an image capturing device is instructed to capture a first image 204 of a target object.
  • such instruction is inputted by a user 202 or issued by a command received from a control unit of the mobile device.
  • Defect inspection algorithms 208 are used to process the image 204 to detect whether the target object has a defect 210. These defect inspection algorithms are guided by different types of visual knowledge 206, which will be elaborated later.
  • a confidence score 215 for the detection on whether the target object has the defect 210 is computed and fed back to the user 202 to assist the user 202 or the control unit to capture one or more images (or snapping one or more images using a camera) with more optimal imaging parameters (angle, distance, etc.) while performing different automatic inspection tasks (e.g. object recognition, scene reconstruction, feature detection and the like).
  • the image capturing device determines, based on the calculated confidence score 215, whether to generate better view recommendations 220 that may include improved imaging parameters with respect to translation and/or rotation movement of the imaging capturing device.
  • the improved imaging parameters may be generated, if required, after considering the calculated confidence score.
  • the image capturing device upon receiving an input, is instructed to capture, according to the improved imaging parameters, a second image of the object that results in calculation of an improved confidence score.
  • better imaging observations are incrementally achieved for each specific inspection task, thereby facilitating better detection performance.
  • the collated confidence scores can be used to fine tune detection of the common defect on the object or another object (of a same type as the object).
  • “Better imaging observations” can refer to better imaging quality achieved through application of improved imaging parameters with respect to translation and/or rotation movement of the imaging capturing device.
  • the calculated confidence score is generated as feedback to the user after comparing the calculated confidence score against a threshold, wherein the threshold is used to reduce a number of false detection.
  • the improved imaging parameters can be generated after comparing the calculated confidence score against the threshold. Thereafter, upon receiving an input, the image capturing device is instructed to capture, according to the improved imaging parameters, a second image of the object that results in calculation of an improved confidence score
  • a computer program product (i.e. application or program) may be provided to cause the execution of the above-described defect inspection algorithm on a program controlled entity.
  • Such computer program product and its instructions thereof may be stored in a memory accessible to a processor and be executable by the processor.
  • the computer program product may be provided or delivered as a recording means, including a non-transitory computer readable medium, such as a memory card, USB-stick, CD-ROM, DVD or also in form of a downloadable file from a server in a network. This can, for example, be achieved through a wireless communication network by the transmission of a respective file having the computer program product.
  • the computer program product includes a mobile application tied to an operating system of a smartphone.
  • a decision-theory (also known as by-contrary decision) based single-view inspection without reference images where an observed geometric pattern might occur in a random image, e.g. the detection of a rare event such as a crack on the inspected surface, is based on decision theory to produce a confidence score that represents defect pattern strength (e.g. crack strength).
  • the inspection may exploit generic knowledge that is learned beforehand in the decision-theory based single-view inspection method.“Generic knowledge” used herein refers to factual information or concepts in computer visioning system for detection of defect(s) in an object.
  • a classification-based single-view inspection with training images where visual features can be used to classify image patches into normal or abnormal (such as non-erosion or erosion regions), based on learned knowledge from training data to produce a confidence score that represents class membership probabilities.
  • a single-view inspection with 2D reference image where missing materials or geometry defects are detected by comparing the real (detected) shape against the ideal shape mapped from a reference image that has no defect to produce a confidence score that represents registration quality (e.g. number of matched feature points after registration).
  • a single-view inspection with 3D reference model is provided where missing materials/geometry defects are detected by comparing the real (detected) shape against the ideal shape registered/projected from a reference 3D model that has no defect to produce a confidence score that represents registration quality (e.g. number of matched feature points after registration).
  • the visioning system comprises a plurality of operation modes for selection, wherein each operation mode is for detecting one type of defect, and the processing of a captured image of an object to detect whether the object has the defect is performed according to the selected operation mode.
  • the plurality of operation modes may include at least two of the following operation modes:
  • the method is conducted based on a“a contrario " statistical framework (also known as‘by contrary decision’).
  • a“a contrario " statistical framework also known as‘by contrary decision’.
  • the occurrence expectation should be small in the image.
  • a geometric pattern is considered meaningful in the “a contrario " framework if expectation of its occurrence would be very small in a random image.
  • the lower the probability of such event the more meaningful the geometric pattern is.
  • the expected occurrence of such event represented by e
  • e can be inputted by a user or predefined by the apparatus.
  • Mathematical expectation is the summation or integration of a possible value from a random variable. It is also known as the product of the probability of an event occurring, denoted as P(x), and a value corresponding with an actual observed occurrence of the event.
  • the expected value is a useful property of any random variable.
  • the mathematical expectation is usually notated as E(X) and such expected value can be computed by overall summation of distinct values that the random variable can take.
  • Figure 3 illustrates an example of a crack on a target object (e.g. engine blade) represented by a discrete line segment c ⁇ ⁇ p, r, a ⁇ , characterized by its starting point p (from hole), its length r e M and its direction a.
  • a target object e.g. engine blade
  • the intensity image is locally normalized by dividing it by its median on a small neighborhood. That is, for q— (x, y), a normalized intensity image is defined as
  • Equation (2) Equation (2)
  • NFA(c ) #(? c (t(c)) ⁇ e - Equation (3)
  • F c t is the probability that a random variable t is larger than a given threshold.
  • the value e corresponds to an expected number of false detections under the hypothesis 7f 0 , where 7f 0 is a stochastic model for unstructured data.
  • NFA (c) is a stochastic model for unstructured data.
  • NFA (c) is a stochastic model for unstructured data.
  • NFA (c) the more meaningful the crack c.
  • a detection method is said to produce e -meaningful detections (events), when the expected number of false detections is bounded by e. See definition of #C in Equation (4) below.
  • the starting point p is set in the w x w window (depicted as square in Figure 3) and centered at a lower vertex o on a major axis 304 of an ellipse 302 representing a hole.
  • the w x w window represents an image patch / of a captured image (e.g. 204 of Figure 2). Details on how an image patch of a captured image can be obtained will discussed later.
  • Oi to 0 5 are points on a boundary of the ellipse 302. In one example, each of these points Oi to 0 5 can be used as an origin to generate fan-shape scanning regions for crack detection.
  • the length of crack r ranges discretely from r min to r max : r e [r min , r max ] .
  • the crack direction a is within a p/2 range from the ellipse orientation: a e [f - ⁇ , f + ⁇ ].
  • the number of discrete direction at a given crack length r is set to be K(r) - ⁇ r . Therefore, the number of all possible cracks in the image patch / is
  • #C is the number of all possible cracks in an image patch /
  • # ⁇ p e J ⁇ f w (o) ⁇ is the number of a probability that starting point p lies on the lower vertex o of the ellipse
  • a crack is detected as e-meaningful in the image patch / if its strength t(c) is larger than the threshold, Ur, e), defined in Equation 6. - Equation (6),
  • FIG. 4A is a test image for detecting cracks according to one example of the present disclosure. On region 400 is marked out as an example.
  • Figure 4B is an exploded view of the region 400 of Figure 4A that is marked by an inspector to indicate a crack in several apertures 402 and it matches with a detection result in Figure 4C that is indicative of whether the test image has a defect.
  • Figure 4C shows computed crack lines 404 in the several apertures 402.
  • a SVM is a discriminative classifier formally defined by a separating hyperplane.
  • the algorithm Given labelled training data (supervised learning), the algorithm outputs an optimal hyperplane which categorizes the training data according to their labels.
  • a SVM model is a representation of the examples (training data) as points in space, mapped so that the examples (training data) of the separate categories are divided by a clear gap that is as wide as possible. New examples (test data) are then mapped into that same space and predicted to belong to a category based on which side of the gap they fall.
  • image patches are classified into normal/abnormal (e.g., erosion or non-erosion) regions based on learned knowledge.
  • confidence score is given by class membership probabilities.
  • An image patch described in the present example refers to a portion (usually a rectangle window) of a training image or a test image that is being processed for detection of a defect.
  • the patches are automatically cropped from the original training image or the original test image with some overlaps between each patch in a sliding-window way.
  • a plurality of image patches e.g.
  • each image patch 601 , 602, 605, 606 and 607) are cropped from a test image 610 such that each image patch has an overlapping portion with at least two image patches proximate to it.
  • the image patch may be highlighted (for example, displaying as a black-outlined box or a hashed-outlined box) on the original test image for viewing.
  • FIG. 5 The flowchart of classification-based single-view inspection with training images is shown in Figure 5.
  • Figure 5 indicates that training data are provided to an image capturing device (visioning system) at a step S502, and the predetermined features (such as colour, texture, and histogram of gradient) are extracted out of each training image obtained from the training data at a step S504.
  • the extracted features of the respective training images are sent to a classifier (e.g. a support vector machine) to learn before grouping each training image accordingly at a step S506.
  • the image capturing device is configured to output and monitor a classification result indicating whether a sample object comprises a defect and/or a number of false detections at a step S512.
  • the classification result may be used to fine-tune the rules for classifying the training images and/or test images.
  • the image capturing device is configured to compute a confidence score 215 from the classification result as a feedback to the image capturing device and/or a user.
  • the image capturing device processes the testing data in a similar fashion as the training images. For instance, testing data are provided to the image capturing device at a step S508, and the predetermined features (such as colour, texture, and histogram of gradient) are extracted out of each captured image obtained from the testing data at a step S510.
  • the image capturing device is also configured to compute a classification result indicating whether the test object comprises a defect at the step S506.
  • the image capturing device is configured to output and monitor a classification result indicating whether the test object comprises a defect and/or a number of false detections at a step S512, and to fine-tune the rules to classify the training images and/or test images.
  • the image capturing device is configured to compute a confidence score 215 from the classification result as a feedback to the image capturing device and/or a user.
  • An example of erosion detection result is shown in Figure 6B.
  • Image patches indicated on a test image 610 as a plurality of black-outlined boxes in Figure 6B are erosion regions 620.
  • Geometry defects on a surface of a material, or missing portions from a surface of a material can be detected quickly from a comparison between a 2D image of the surface of the material with a single 2D reference image that has no geometry defect.
  • This is referred to as Single-view inspection with 2D reference image in the present disclosure.
  • the discussed technique applies to a new image and a reference image that is mono-chromatic or grayscale (instead of a full colour image).
  • a rigid transform between a reference image 702 with ideal geometry and a new image 704 is determined by registration of their corresponding key point positions (e.g. edges, corners and the like).
  • image registration is a process of transforming different sets of data into one coordinate system (space). Data may be multiple images, or images captured from different times, viewpoints or sensor.
  • the ideal geometry 706 in the reference image 702 is transformed to a coordinate space of the real geometry 708 and missing materials/geometry defects is detected by comparing features at the corresponding positions.
  • a confidence score 215 is obtained based on registration quality (e.g., number of matched feature points) as a feedback. Thereafter, better observation (e.g., translation/rotation of hand-held imaging devices) recommendation is given according to the obtained scores and/or score changes over time as more images are captured for capturing a defect.
  • registration quality e.g., number of matched feature points
  • better observation e.g., translation/rotation of hand-held imaging devices
  • Geometry defects on a surface of a material, or missing portions from a surface of a material can also be detected by the vision system with a 3D reference model 802 that has no geometry defect.
  • the flowchart of single-view inspection with 3D reference model is shown in Figure 8.
  • a new 2D image 804 is registered to the 3D model 802 by registration of their corresponding key point positions (typically represented as n). With sufficient number of corresponding points (i.e. n > 3), the optimal projective transformation, associated viewing angles and other registration parameters can be derived. This means that the registration will use optimal viewing parameters derived from the corresponded key points. Then, with the optimal registration parameters, the 3D model is projected to a (2D) projection view 806 that corresponds to the respective 2D view of the new image 804.
  • Efficient Perspective- n- Point (EPnP) method developed by Lepetit et al. in their 2008 International journal paper can be used to estimate the 2D projection view 806 that corresponds to the respective 2D view of the new image.
  • Missing materials and/or geometry defects can be detected by comparing the corresponding positions between a real geometry 808 detected from the new 2D image 804 and an obtained ideal geometry of the projection view 806 in a coordinate space determined by transformation.
  • a confidence score 215 is calculated based on registration quality (e.g., number of matched feature points). Thereafter, better observation recommendation (e.g. translation and/or rotation movement of an imaging capturing unit, a sensor, a camera or the like) is calculated and generated according to the confidence scores/score changes.
  • registration quality e.g., number of matched feature points.
  • better observation recommendation e.g. translation and/or rotation movement of an imaging capturing unit, a sensor, a camera or the like
  • the techniques disclosed herein can also be used in imaging sensor planning in robotic vision.
  • the objective of imaging sensor planning in robotic vision can be summarized as follows:
  • a sensor planning system 900 is able to develop strategies to automatically determine sensor parameter values 910 that can achieve a task with a certain degree of satisfaction. With such strategies, sensor parameters values 910 can be selected and can be purposefully changed in order to effectively perform the task at hand. Sensory systems are then able to operate more flexibly, autonomously, and reliably. The process of sensor planning system is plotted in Figure. 9A.
  • Figure 9B One existing model of sensor planning that is discussed in International Patent Application No. PCT/SG2017/050175 is illustrated in Figure 9B.
  • the model in Figure 9B comprises three parts: system setup 952 for different types of objects, task specification 954 for objects of the same type and online inspection 956 for each object.
  • the individual steps for inspection 956 are not relevant to the present disclosure and are not elaborated herein. It should be appreciated that various sensing and inspection steps can be used for the online inspection 956 in the existing method of sensor planning.
  • the system setup 952 consists of the following two parts:
  • Sensor calibration Calibrate different sensors (texture sensor, geometry sensor) into a same coordinate system, i.e. the robotic base coordinate system.
  • Geometry sensor Field of view, clear distance, depth of field, precision for each measurement point, point density at different sensing distance.
  • Texture sensor Field of view, Depth of field
  • the task specification 954 seeks to specify global parameters 960 for objects of the same type, which includes the desired geometry and texture quality.
  • An object model 958 for the objects to be inspected has to be prepared. If sensors and the robot (the actuator) are fixed for inspecting different types of defects on different objects (like in examples of the present disclosure), system configuration remains unchanged, and thus, the sensors and the robot only need to be calibrated once. For inspecting objects of a same type (like in examples of the present disclosure), the task specification can be performed once. If new training data is available, prior knowledge training can be updated to include the new training data.
  • the criteria for optimal sensor setup in existing sensor planning methods is geometry- oriented (which aims at optimal output geometry quality).
  • the criterion for the techniques disclosed herein is defect-oriented which aims at better detection rate.
  • existing sensor planning includes object recognition and feature detection (which is similar to (at least) one of the tasks implemented through the techniques disclosed herein), existing sensor planning is reliant on low level sensing parameters such as resolution during the actual planning process.
  • the techniques disclosed herein focus on defect-level pattern recognition tasks and the corresponding detected result is actually returned as a feedback during the imaging process.
  • the input to the visual system can be generic knowledge, training image data or reference images.
  • These different operation modes are non-model based (i.e. no sensor model and/or object model are used), and such non-model-based approaches are used to determine the best next view and sensor settings to incrementally acquire the object information.
  • This is different from conventional sensor planning methods, because these methods use model- based approaches that require sensor model and object model to determine the optimal sensor placements and a shortest path through these viewpoints.
  • Defect-oriented confidence scores are used to guide the inspector (user) to find the optimal imaging parameters and it is applicable to generic visual inspection of objects, structures & architectures. There is no requirement for the sensor set-up to be optimized before detection of defects.
  • the defect-oriented confidence scores are designed for defect-level pattern recognition tasks.
  • existing sensor planning methods are geometry-oriented and have to ensure that captured images have unobstructed view, are in focus, are properly magnified, are well-centered within a field-of-view, and imaged with sufficient contrast.
  • a toolbox based on the proposed solution is useful for industries that require defect inspection with mobile (especially hand-held) imaging devices.
  • the applications may include:
  • classification-based single-view inspection with training images can be adapted for detecting other surface defects, such as, discoloration and burning, by using a classifier that is trained by a new set of training images of relevant defects.
  • a method, apparatus and/or a (visual or vision) system for visual inspection with active 3D measurement is also provided.
  • Such method, apparatus and/or a (visual or vision) system is referred herein as Snap2lnspect3D.
  • Snap2lnspect3D Once defects in 2D images are detected, for instance, through the snap2inspect or snap2inspect2D techniques disclosed previously, the targeted defect regions can be measured actively with an adaptive illumination pattern.
  • This exemplary technique provides key 3D measurements such as geometry of the inspected surface in absolute scale and profile data of defects (length, width and depth) while keeping the number of image acquisitions low or within to a few snaps. It should be appreciated that this exemplary technique provides a lower cost, and yet faster method to obtain accurate 3D measurement of a detected defect as compared to existing methods.
  • correspondence problem One limitation in 3D reconstruction is known as correspondence problem, particularly when observing non-textured objects.
  • methods based on structured light can be used to create correspondences and give specific code word to every position on an image.
  • a structured light pattern is projected onto a scene and imposes illusion of texture on an object, thereby increasing the number of correspondence on non-textured objects. Therefore, surface reconstruction is possible when looking for differences between projected and recorded patterns.
  • Coded structured light is preferred because each element of the pattern can be unambiguously identified on the images.
  • the aim of coded structured light is to robustly obtain a large set of correspondences per image independently of the appearance of the object being illuminated and the ambient lighting conditions.
  • a set of patterns may be successively projected onto a surface.
  • the codification is temporal.
  • Such kind of patterns can achieve high accuracy in measurements, but the ability to measure moving surfaces is restricted to one-shot scans (i.e. not robust to motion acquisition).
  • classical-based multiplexing techniques like M-array based techniques perform one-shot 3-D reconstruction with good accuracy results, these techniques produce sparse (feature wise) reconstructions.
  • the following example demonstrates that by targeting at defect regions, it is possible to perform motion-robust and accurate 3D measurement of targeted defect using a hand-held device.
  • FIG. 10 The workflow of the visual inspection with active 3D measurement, or Snap2lnspect3D is shown in Figure 10.
  • an image is captured (S1 ) using an image capturing unit of a portable imaging device (e.g. a camera on a smartphone/tablet).
  • a portable imaging device e.g. a camera on a smartphone/tablet.
  • a spatial coded pattern is used to reconstruct (S2) sparse 3D points of an underlying surface of a detected defect region within a single snap (or by capturing a single image of the defect region).
  • the image captured by the camera can be overlaid with the surface as texture and the absolute scale of the defect on the image can be inferred.
  • the system projects an adaptive pattern to measure the target region of the defect (i.e. a defect region) for dense and accurate 3D profiling (S4).
  • a light projector upon detection that an object has a defect, for instance, by the snap2inspect3D techniques, a light projector is instructed to project a coded structured pattern on the object. Thereafter, the image capturing device is instructed to capture a two dimensional image of the object comprising the projected coded structured pattern.
  • a three dimensional (3D) model of the object is constructed from the captured two dimensional image to obtain surface geometry of the object to facilitate 3D measurements of the defect.
  • the coded structured pattern is a monochromatic spatially coded pattern.
  • FIG. 1 1 A hardware set-up of a proposed vision system 1 100 is shown in Figure 1 1 . It comprises of an image capturing device 1 105, such as in the present example, a camera of a smart phone, and a portable projector 1 1 10. To set up the vision system 1 100, the image capturing device 1 105 is mounted on an adjustable clamp (not shown in Figure 1 1 ) together with the projector 11 10 at a certain angle q between the imaging capturing device 1 105 and a projected pattern 1 1 15 based on a target object 1 120 to be measured.
  • a cable may be used to connect between an input/output port (e.g. USB slot) on the image capturing device 1 105 and an input/output port (e.g.
  • the active device is typically a digital light projector and is modelled as the inverse of a camera.
  • a camera When a camera is used, light from the environment is focused on an image plane and captured. Such process reduces the dimensions of the data taken in by the camera from three dimensions to two dimensions (i.e. light from a 3D scene is stored on a 2D image), and hence a system calibration is required before dimensions of a defect on a 3D object can be measured from the camera images.
  • Figure 12 shows coordinate systems to be considered for the projector 1 1 10 and the camera 1 105 for system calibration.
  • the dimensions of a three-dimensional (3D) point P (X w , Y w , Z w ) taken in by a camera 1 105 is reduced from three dimensions to two dimensions, and represented as p (u c ,v c ).
  • the coordinates of the point P is computed with respect to a 3D coordinate space 1202 formed by three axes X w ,Y W , Z w , which intersect at an origin O.
  • the three axes X w , Y w , Z w are orthogonal to one another.
  • the coordinates of p (u c ,v c ) are formed with respect to a 2D camera coordinate space formed by two axes u and v, which intersect at an origin o.
  • the origin o of the 2D camera coordinate space corresponds to an origin O c of a camera coordinate frame formed by three axes X c , Y c , Z c .
  • the three axes X c , Y c , Z c are orthogonal to one another.
  • the 3D point P has a corresponding projecting point p’ (u p , v p ) at the projector 1 1 10.
  • the coordinates of p’ (u p , v p ) are formed with respect to a 2D projector coordinate space formed by two axes u and v, which intersect at an origin o.
  • the origin o of the 2D projector coordinate space corresponds to an origin O p of a projector coordinate frame formed by three axes X p , Y p , Z p .
  • the three axes X p , Y p , Z p are orthogonal to one another.
  • intrinsic parameters of the camera 1 105 and the projector model 1 1 10, and extrinsic parameters between the projector 1 110 and the camera 1 105 are determined.
  • the terms“camera” and“camera model” and the terms“projector” and “projector model” are used interchangeably in the present example.
  • the intrinsic parameters of the camera model 1 105 include a focal length and co-ordinates of a principal point, which describes a relationship between a 3-D point P (X w , Y w , Z w ) and a corresponding projection onto an image plane.
  • the extrinsic parameters include rotation and translation between the coordinate systems of the camera and the projector, which describes the relative position and orientation between them.
  • calibration can be performed by using a black image with an illuminated pixel as a pattern.
  • only one point can be reconstructed through triangulation using pixel coordinates of the illuminated pixel in the pattern and the corresponding coordinates in the camera image.
  • the intrinsic parameters of the camera model 1 105 can be estimated by widely used calibration processes, such as Direct linear transformation (DLT) method, Zhang’s method and the like.
  • DLT Direct linear transformation
  • the system may use a method of projecting a calibration pattern such as through the use of a planar checkerboard to minimize calibration complexity and cost under a fixed setting with the camera 1 105 and the projector 11 10 fixed.
  • An example of the planar checkerboard is shown in Figure 13d.
  • the planar checkerboard as shown in Figure 13d can be replaced with asymmetrical circle patterns as shown in Figure 13e for the ease of real-time detection.
  • the proposed system 1 100 and the corresponding method discussed above consider the projector as an inverse camera which maps 2D image intensities into 3D rays, and thus make the calibration of a projector-camera system a similar procedure as that of stereo cameras. In this way, by having the 2D projected points and its 3D correspondences, the system 1 100 can be calibrated using any standard stereo camera calibration method.
  • a monochromatic spatially coded pattern is preferred. Colored patterns can also be used although colored patterns are generally less robust than monochromatic patterns.
  • the coded pattern can be adapted based on an M-array based theory, with uniqueness of the code for each 3 by 3 window.
  • An algorithm such as that is proposed by Albitar et al; IEEE int. conf. on image processing, pages 529-532, 2007 (hereinafter“Albitar et al”) may be used.
  • An alphabet of three symbols that is associated to geometrical shapes (disc, circle, and stripe) may be used.
  • the length of the code associated to each primitive is 9 since a 3x3 window centred on this primitive is considered in this example.
  • an obtained matrix of dimensions 27x29 which means 783 primitives, verifying first three desired criteria and using three symbols is obtained through the algorithm applied (see Figure 13a).
  • the first three desired criteria refers to the three desired criteria used in the algorithm proposed by Albitar et al to ensure that the system decodes the coded patterns correctly are central symmetry, uniqueness of the code of each 3x3 window, and hamming distance being more than 3 (i.e. each code word is different from each other by at least three symbols).
  • Projected pattern needs to be segmented and decoded for 3D reconstruction (see Figures 13b and 13c).
  • the contours are segmented and then classified into three pattern primitives. Once each primitive together with its neighbors is detected, its code is determined. Finally, the 3D position of each decoded point can be calculated through triangulation between the observed pattern and the original one.
  • the projector 1 1 10 can be configured to project a structured coded pattern on a 2D image captured by the camera 1105 for defect detection. Thereafter, a 3D model is reconstructed by decoding the projected coded pattern as discussed to obtain 3D coordinates to detect a defect on the 2D image. Once a defect is identified on the 2D image with its underlying surface parameters estimated using the method described above, the 3D coordinates of any points in the defect region can be calculated from the 2D image.
  • the system 1 100 is configured to generate a defect-specified pattern to measure the defect automatically.
  • Figure 14 shows essentially the same elements as Figure 12 but in addition, there is shown a cube 1404 provided for illustrative purposes and it represents the 3D object to be scanned for defects.
  • the system 1 100 is configured to assume a point, p, as a centre of defect in a 2D image, with its 3D point, P (X W ,Y W ,Z W ), in a camera frame 1402 (i.e. video frame captured by the camera 1 105). Based on the camera-projector calibration,
  • K is the intrinsic parameters for the projector
  • T is the extrinsic parameter between camera frame 1402 and projector frame 1406.
  • Various pattern codification strategies can be applied. For example, with reference to Figure 17, to measure a depth of a dent on a surface, a pattern of four-points are required, where three points form a reference plane with the remaining one point as target point. Thus, the depth of the dent can be calculated as distance between the target point and the reference plane.
  • system calibration should be performed.
  • a user may point both the camera 1 105 and projector 1 1 10 toward a calibration board with asymmetrical circle patterns (e.g. as shown in Figure 13e).
  • the detected patterns are automatically marked on a screen 1502 by the system 1 100 is illustrated in Figure 15, and recorded for correlation.
  • the calibration is repeated at least a further two times, with the camera 1 105 and the projector (of the system 1 100) 1 1 10 tilted at a different angle Q toward the calibration board with asymmetrical circle patterns to collect at least three sets of data.
  • the system calibration result with 10 sets of data is illustrated in Figure 16.
  • the image capturing device 1 105 and the light projector 1 1 10 are directed towards a calibration board comprising an area 1510 with calibration pattern 1506 such that the light projector 1 1 10 projects asymmetrical circle patterns 1504 outside the area 1510 with calibration pattern 1506.
  • the calibration pattern 1506 and the projected asymmetrical circle patterns 1504 are detected by capturing, using the image capturing device 1 105, an image of the calibration board comprising the calibration pattern 1506 and the projected asymmetrical circle patterns 1504 at a first instance.
  • the detection of the calibration pattern 1506 and the projected asymmetrical circle patterns 1504 is graphically represented as lines 1512 and 1508 respectively,
  • the line 1508 is formed by joining each center point of each asymmetrical circle of the pattern 1504 together
  • the line 1512 is formed by joining each center point of each asymmetrical circle of the pattern 1506 together.
  • the calibration pattern 1506 and the projected asymmetrical circle patterns 1504 are detected again for each additional image captured at the different angle q by the image capturing device 1 105. After capturing at least three images at respective different angles, a calibration result is obtained based on collated data of the detected calibration pattern and detected projected asymmetrical circle patterns.
  • the calibration pattern 1506 may be different or the same as the projected asymmetrical circle patterns 1504.
  • the calibration pattern 1506 can be asymmetrical circle patterns or planar checkerboard pattern.
  • the calibration pattern 1506 in the area 1510 may be overlaid with the structured coded pattern projected by the light projector 1 110 during calibration.
  • the system 1 100 can perform dimensional measurement using stereo-triangulation on the selected object. For example, a height of an object on a surface, or a depth of a cavity in a surface can be measured.
  • Figure 17 illustrates an example of height measurement of a rectangular object 1704 on a surface 1706.
  • Four illuminated dots (or points) 1702 are projected onto the object 1704 placed on the surface 1706, and their 3D coordinates 1708 with respect to a centre of the camera 1 105 is calculated in real-time based on the triangulation.
  • Height or depth on the upper left corner is calculated by the distance from a centre point 1710 to the plane formed by the rest of three points, for example, (x 0 , y 0 , z 0 ), (x 2 , y 2 , z 2 ), and (x 3 , y 3 , z 3 ).
  • the four illuminated dots 1702 are projected onto a surface containing an elevated or a dented point for measuring respective height or depth of the point relative to the surface, wherein three of the four illuminated dots 1702 for forming a reference plane are projected on the surface and the remaining one illuminated dot (e.g.
  • the centre dot or point 1710 in Figure 17 is projected on the point.
  • the respective height or depth of the point relative to the surface based on distance between the point and the reference plane can be calculated.
  • a confidence score is provided to assist an inspector to capture images using optimal imaging parameters (angle, distance, etc.) when performing different automatic inspection tasks.
  • a flow chart 1800 in Figure 18 describes how the proposed method performed through the corresponding system 1 100 fits into the framework of Snap2lnspect3D.
  • an inspector 1802 obtains an image of a target object and performs a one shot-scan for 3D surface reconstruction. Thereafter, Snap2lnspect3D technique 1804 described above is applied to detect a defect of the target object using visual knowledge and surface geometry involving 3D measurement (1806). That is, Snap2lnspect3D 1804 involves active sensing for surface geometry and 3D defect measurement (1808). In one example, four points may be illuminated on a detected defect region, and a 3D measurement of a defect can be subsequently made if positions of the projector 1 1 10 of Figure 1 1 and camera 1 105 of Figure 1 1 are fixed.
  • a confidence score 1810 for the detection on whether the object has a defect is computed and fed back to an inspector 1802 to assist the inspector 1802 to capture one or more images with more optimal imaging parameters (angle, distance, etc.).
  • the system 1 100 determines, based on the calculated confidence score 1810, whether to generate better view recommendations 1814.
  • better view recommendations 1814 may include improved imaging parameters with respect to translation and/or rotation movement of the camera 1 105 and generates the improved imaging parameters if required.
  • the camera 1 105 may be instructed to apply the improved imaging parameters to capture a second image of the object that results in calculation of an improved confidence score.
  • the proposed method and corresponding system 1 100 of Figure 1 1 for visual inspection with active 3D measurement offers active defect measurement guided by defect detection, which could be readily implemented using existing hardware, such as smartphone or tablet with image capturing capabilities, and a hand-held projector into a low-cost palm-sized platform for on-site visual inspection.
  • the proposed method and corresponding system 1 100 further allows active 3D measurement with adaptive illumination pattern guided by the image-based defect detection. It provides key 3D measurements such as geometry of the inspected surface in absolute scale and profile of defects (length, width and depth) while keeping the scanning time within a few snaps.
  • the proposed method and corresponding system 1 100 advantageously provides intelligent defect detection with absolute scale, where the image captured by the camera 1 105 can be overlaid with the surface geometry as texture. Based on the underlying geometry, the absolute scale of the detected area on the image can be inferred. Thus, task-specified dimensional criteria such as the maximum acceptable width of crack can be taken into consideration during defect detection and decision making process.
  • the image capturing device 1 105 in Figure 1 1 may be a computing or mobile device, for example, smart phones, tablet devices, and other handheld devices.
  • the image capturing device 1 105 in Figure 1 1 and the projector 1 1 10 in Figure 1 1 may be able to communicate through other communications network, such as, wired network, but these are omitted from Figure 1 1 for the sake of clarity.
  • the architecture of a system or an apparatus proposed in an example of the present disclosure may be an apparatus 1902 in Figure 19 and comprises a number of individual components including, but not limited to, processing unit 1916 (or processor), a memory 1918 (e.g. a volatile memory such as a Random Access Memory (RAM) for the loading of executable instructions 1920, the executable instructions defining the functionality the apparatus 1902 carries out under control of the processing unit 1916.
  • the apparatus 1902 also comprises a network module 1925 allowing the apparatus 1902 to communicate over the communications network 1908 (for example the internet, Bluetooth network).
  • User interface 1924 is provided for user interaction and may comprise, for example, conventional computing peripheral devices such as display monitors, computer keyboards and the like.
  • the apparatus 1902 may also comprise a database 1926 to store defect detection results (and corresponding confidence scores and/or system settings of the image capturing device 1905 as well as the projector 1 1 10).
  • the database 1926 may also be configured to store a coded projection pattern for projecting on a defect region of an object. It should also be appreciated that the database 1926 may not be local to the apparatus 1902.
  • the database 1926 may be a cloud database.
  • the processing unit 1916 is connected to input/output devices such as a computer mouse, keyboard/keypad, a display, headphones or microphones, a video camera and the like (not illustrated in Figure) via Input/Output (I/O) interfaces 1922.
  • input/output devices such as a computer mouse, keyboard/keypad, a display, headphones or microphones, a video camera and the like (not illustrated in Figure) via Input/Output (I/O) interfaces 1922.
  • I/O Input/Output
  • the components of the processing unit 1916 typically communicate via an interconnected bus (not illustrated in Figure 19) and in a manner known to the person skilled in the relevant art.
  • the processing unit 1916 may be connected to the network 1908, for instance, the Internet, via a suitable transceiver device (i.e. a network interface) or a suitable wireless transceiver, to enable access to e.g. the Internet or other network systems such as a wired Local Area Network (LAN) or Wide Area Network (WAN).
  • the processing unit 1916 of the apparatus 1902 may also be connected to one or more external wireless communication enabled device 1904 through the respective communication links 1910 and 1912 via the suitable wireless transceiver device e.g. a WiFi transceiver, Bluetooth module, Mobile telecommunication transceiver suitable for Global System for Mobile Communication (GSM), 3G, 3.5G, 4G telecommunication systems, or the like.
  • GSM Global System for Mobile Communication
  • the architecture of a system or an apparatus proposed in an example of the present disclosure may be a device 1904 in Figure 19.
  • the device 1904 may comprise a number of individual components including, but not limited to, microprocessor 1928 (or processor), a memory 1930 (e.g. a volatile memory such as a RAM) for the loading of executable instructions 1932, the executable instructions defining the functionality the device 1904 carries out under control of the microprocessor 1928.
  • the device 1904 also comprises a network module (not illustrated in Figure) allowing the device 1904 to communicate over the communications network 1908.
  • User interface 1936 is provided for user interaction and control that may be in the form of a touch panel display and presence of a keypad as is prevalent in many smart phone and other handheld devices.
  • the device 1904 may also comprise a database (not illustrated in Figure), which may not be local to the device 1904 but a cloud database.
  • the device 1904 may include a number of other Input/Output (I/O) interfaces 1934 as well but they may be for connection with headphones or microphones or projectors, Subscriber identity module (SIM) card, flash memory card, USB based device, and the like, which are more for mobile device usage.
  • I/O Input/Output
  • SIM Subscriber identity module
  • the software and one or more computer programs stored may include, for example, applications that may include one or more applications for e.g. internet accessibility, operating the device 1904 and apparatus 1902 (i.e. operating system), network security, file accessibility, database management, which are applications typically equipped on a desktop or portable (mobile) device.
  • the software and one or more computer programs may be supplied to a user of the device 1904 and/or the apparatus 1902 encoded on a data storage medium such as a CD-ROM, on a flash memory carrier or a Flard Disk Drive, and are to be read using a corresponding data storage medium drive for instance, a data storage device (not illustrated in Figure 19).
  • Such application programs may also be downloaded from the network 1908.
  • the application programs are read and controlled in its execution by the processing unit 1916 or microprocessor 1928. Intermediate storage of program data may be accomplished using RAM 1920 or 1930.
  • the computer programs may be stored on any machine or computer readable medium that may be non-transitory in nature.
  • the computer readable medium may include storage devices such as magnetic or optical disks, memory chips, or other storage devices suitable for interfacing with a general purpose computer or mobile device.
  • the machine or computer readable medium may also include a hard-wired medium such as exemplified in the Internet system, or wireless medium such as exemplified in the Wireless LAN (WLAN) system.
  • WLAN Wireless LAN
  • the computer program when loaded and executed on such a general-purpose computer effectively results in an apparatus (e.g. 1 105 and/or 1 1 10 in the respective Figures) that implements the steps of the computing methods in examples herein described.
  • examples of the present disclosure may have the following features.
  • a method for detecting a defect on an object comprising:
  • the method may further comprise:
  • the method may further comprise:
  • each operation mode is for detecting one type of defect
  • the plurality of operation modes includes at least two of the following operation modes:
  • an operation mode to detect a crack on the object using a decision-theory based algorithm an operation mode to detect erosion on the object using a classification algorithm that trains on a plurality of training images
  • an operation mode to detect missing material on the object using an algorithm that matches features points of a three dimensional (3D) model with feature points of a captured image of the object.
  • the calculated confidence score may generated as feedback to the user after comparing the calculated confidence score against a threshold, wherein the threshold is used to reduce a number of false detection.
  • the method may further comprise:
  • the coded structured pattern may be a monochromatic spatially coded pattern.
  • the method may further comprise: adapting the coded structured pattern based on viewpoint between a defect and the image capturing device; and/or
  • tuning colours of the coded structured pattern in regard to a specific defect region and/or increasing spatial sample resolution for a target defect region considered.
  • the method may further comprise:
  • the calibration pattern may be a plurality of asymmetrical circle patterns.
  • the method may further comprise:
  • the method may further comprise:
  • An apparatus for detecting a defect on an object the apparatus comprises:
  • a processor configured to execute instructions in a memory to control the apparatus to:
  • an image capturing device e.g.1 105 of Figure 1 1
  • a first image e.g. 204 of Figure 2
  • an object e.g. 1704 of Figure 17
  • a confidence score (e.g. 215 of Figure 2) for the detection on whether the object has the defect
  • the calculated confidence score as a feedback to a user (e.g. 202 of Figure 2); determine based on the calculated confidence score whether to generate improved imaging parameters (e.g. 220 of Figure 2) with respect to translation and/or rotation movement of the image capturing device;
  • the apparatus may be controllable to:
  • the apparatus may be controllable to select an operation mode from a plurality of operation modes so that the processing of the captured first image to detect whether the object has the defect is performed according to the selected operation mode,
  • each operation mode is for detecting one type of defect
  • the plurality of operation modes includes at least two of the following operation modes:
  • an operation mode to detect a crack on the object using a decision-theory based algorithm an operation mode to detect erosion on the object using a classification algorithm that trains on a plurality of training images an operation mode to detect missing material on the object using an algorithm that matches features points of a reference two dimensional (2D) image with feature points of a captured image of the object;
  • an operation mode to detect missing material on the object using an algorithm that matches features points of a three dimensional (3D) model with feature points of a captured image of the object.
  • the calculated confidence score may be generated as feedback to the user after comparing the calculated confidence score against a threshold, wherein the threshold is used to reduce a number of false detection.
  • the apparatus may be controllable to:
  • a light projector e.g. 1 1 10 of Figure 1 1 .
  • the coded structured pattern may be a monochromatic spatially coded pattern (e.g. Figure
  • the apparatus may be controllable to:
  • the apparatus may be controllable to:
  • the apparatus may be controllable to:
  • the calibration pattern may be a plurality of asymmetrical circle patterns.
  • the apparatus may be controllable to:
  • the apparatus may be controllable to:
  • the apparatus is controllable to: calculate height or depth of the point relative to the surface based on distance between the point and the reference plane.

Abstract

A method and apparatus for visual inspection involving detecting a defect on an object, the method comprising: instructing an image capturing device to capture a first image of an object; processing the captured first image to detect whether the object has a defect; calculating a confidence score for the detection on whether the object has the defect; generating the calculated confidence score as a feedback to a user; and determining based on the calculated confidence score whether to generate improved imaging parameters with respect to translation and/or rotation movement of the image capturing device; generating the improved imaging parameters if required after considering the calculated confidence score; and receiving an input to instruct the image capturing device to capture, according to the improved imaging parameters, a second image of the object that results in calculation of an improved confidence score.

Description

Method For Visual Inspection And Apparatus Thereof
Field
The invention relates to a method for visual inspection and an apparatus thereof, in particular, the visual inspection involves detecting a defect on an object.
Background
Computer vision is a sub field of artificial intelligence that investigate how to make computer algorithms perceive and understand the world of a human through images. During an emulation, light is detected by cameras and/or sensors, while high-level of image understanding is processed by using computer algorithms. There are many applications where computer vision plays an important role. For instance, computer vision can be used for range sensing, industrial inspection of manufactured parts and/or parts that are currently in use and susceptible to wear and tear, as well as object recognition.
Visual inspection of surfaces for defects can cover both geometry (dimension) inspection and surface (texture) inspection of the surfaces. Dimension inspection is based on geometric shapes and typically captured using a positioning device holding a geometry sensor (a depth sensor, or a 3D scanner and etc.), while surface (texture) inspection is based on images. Sensor planning, i.e. identification of sensor locations to capture different parts of an object with required specifications is important for acquiring desired geometric shapes for dimension inspection. On the other hand, surface inspection relies on the imaging quality of camera(s) used. Different views of a target region may need to be acquired for analysis to identify the existence of a defect.
While some inspection systems can capture both 3D geometry and 2D images to enable the dimension inspection and surface inspection, the viewpoints for all these systems are predefined before the sensing, e.g. with active sensor planning. For example, in surface inspection of objects of the same type, manual selection is required to define viewpoints for image capturing, while in dimension inspection systems, CAD-based active sensing is adopted to derive viewpoints for capturing the 3D geometry. Multi-view based inspection systems may acquire inspection data by combining both shapes and images from different viewpoints.
In addition, computer vision is particularly valuable in 3D construction and/or reconstruction because computer vision can be used to obtain a three-dimensional position of an object point from its projective points on the image plane. A 3D construction and/or reconstruction is necessary in applications where accuracy of the 3D shape represents a fundamental factor to the performance of a computer vision device.
In conventional methods, high-end 3D scanners are required to achieve a good 3D reconstruction of a target object under inspection. There are also strict requirements to the positioning of the target object and the quality of images captured for the target object. This means that the set up of the scanner has to be precise and it will take a lot of time and effort to set up before different type of defects can be detected and/or a different type of object can be inspected.
Summary
According to an example of the present disclosure, there are provided a method and apparatus as claimed in the independent claims. Some optional features are defined in the dependent claims.
Brief Description of Drawings
In the drawings, the same reference numerals generally relate to the same parts throughout different views, unless otherwise specified. Embodiments of the invention will be better understood and readily apparent to one skilled in the art from the following written description, by way of example only and in conjunction with the drawings briefly described as follows.
Figure 1 A shows how a crack is observed in a hash-outlined box and how a crack is occluded in a black-outlined box in one captured photograph.
Figure 1 B shows how a crack is occluded in the hash-outlined box of Figure 1 A in another captured photograph.
Figure 1 C shows how a crack is observed in the black-outlined box of Figure 1 A in another captured photograph.
Figure 2 shows a flow chart of how a method for performing single-view mobile inspection of defects can be provided according to one example of the present disclosure. Figure 3 is a geometry illustration for detection of a hole crack according to one example of the present disclosure.
Figure 4A shows a test image comprising a crack region that is marked by a visual inspector.
Figure 4B is an exploded view of the crack region in Figure 4A.
Figure 4C shows a plurality of detected cracks in the test image according to one example of the present disclosure.
Figure 5 shows a flow chart of how classification-based single-view inspection is learnt from training images according to one example of the present disclosure.
Figure 6A shows how a plurality of image patches are obtained from a captured image according to an example of the present disclosure.
Figure 6B shows how erosion/corrosion regions can be indicated as on a captured image according to one example of the present disclosure.
Figure 7 shows a flow chart for single-view inspection trained with a 2D reference image according to one example of the present disclosure.
Figure 8 shows a flow chart for single-view inspection trained with a 3D reference model according to one example of the present disclosure.
Figure 9A shows a sensor planning system for computer vision according to one example of the present disclosure.
Figure 9B shows a comparative example of an existing model of sensor planning system for computer vision.
Figure 10 shows a work flow of a proposed Snap-to-inspect three Dimensional system (Snap2lnspect3D) according to one example of the present disclosure.
Figure 1 1 shows a hardware set up for capturing an image for visual inspection with active 3D measurement according to one example of the present disclosure.
Figure 12 shows how system calibration can be formulated according to one example of the present disclosure.
Figure 13A shows an example of a coded pattern according to one example of the present disclosure. Figure 13B shows how a portion of the coded pattern in Figure 13A can be decoded.
Figure 13C shows how a cylinder is reconstructed after decoding the coded pattern in Figure 13A. Figure 13D show an example of a checkerboard pattern that can be used for system calibration according to one example of the present disclosure.
Figure 13E show an example of asymmetrical circle patterns that can be used for system calibration according to one example of the present disclosure.
Figure 14 shows how a projection pattern can be generated for defect measurement according to one example of the present disclosure.
Figure 15 is a screenshot of system calibration with asymmetrical circle patterns according to one example of the present disclosure.
Figure 16 is an example of a system calibration result after system calibration in Figure 15.
Figure 17 is an example of height measurement of a rectangular object on a surface using Snap2lnspect3D.
Figure 18 is a flow chart on how Snap2lnspect3D operates according to one example of the present disclosure.
Figure 19 is a block diagram illustrating a system architecture according to an example of the present disclosure.
Detailed Description
Visual inspections expect quick and accurate evaluation of visual contents for significant abnormities, such as defects, on a 3D object surface. Occlusions and/or shadings may hide defect features from adequate observations and detection. Figures 1 A to 1 C are images of a blade surface taken from different views, and these figures illustrate how occlusion and/or shading can hide defect features. For example, the crack in hashed-outlined box 101 can be observed in Figure 1 A, but hardly seen in Figure 1 B, and the crack in black-outlined box 102 can be clearly observed in Figure 1 C, but is very weak in Figure 1 A. Flence, it is important to use an appropriate view for optimal defect detection.
Mobile (portable or robotic) imaging devices enable flexible observation settings (angles, distances, etc.). A quick snap or image capture of the object surface without intentional planning using a mobile device is likely to result in a sub-optimal observation of defects. Flandheld devices (such as a smart phone or an imaging device that can be held by a user for capturing images) are even more likely to generate sub-optimal observation due to inaccurate positioning. As a result, the picture(s) taken by a handheld device may not be at an optimal viewing angle and distance (or optimal observation settings) to distinguish an abnormity from the background or other content on a target object. This technique imposes challenges to the demand of excellent feature representation and excellent detection algorithms which can detect abnormities under various imaging and lighting conditions.
A further challenge is resolving the scale of the target object where visual inspection tasks require measurements in absolute scale.
A solution for mobile inspection of defects assisted by a variety (or different types) of visual knowledge is provided to address the challenges described above.
“Single view inspection” used herein refers to inspection using only a single view for capturing three dimensional (3D) geometry and/or two dimensional (2D) images to enable dimension inspection and surface texture inspection. This can be referred as“one-shot scan” in the present disclosure.
“Portable” used herein refers to a lighter and smaller version (than usual) of an object that allows it to be easily carried or moved or transported around from one place to another place by a user.
“Engine blade” used herein may refer to a turbine blade of a gas turbine or a steam turbine, or may refer to fan blades or propellers used in an automotive vehicle (such as a jet), industrial condensers, industrial fans and the like.
In one exemplary technique, there is provided a method and a device for single-view mobile inspection of defects assisted by different types of visual knowledge. In one aspect of the exemplary technique, visual knowledge about the defect characteristics, such as training images or reference image, is acquired as part of input of the defect detection algorithms. In another aspect of the exemplary technique, defects are detected from individual single-view images captured by an image capturing device, preferably hand held by a user conducting the inspection, and concurrently, a confidence score is calculated, and returned as feedback to the user. In another aspect of the exemplary technique, better inspection results are incrementally achieved when imaging parameters are tuned, preferably by the user.
In another exemplary technique, there is provided a method, (vision) system and/or apparatus for performing single-view mobile inspection of defects that is guided by different types of visual knowledge. These method, system and/or apparatus are termed in the present disclosure as“snap- to-inspect”, “snap2inspect2D” or “snap2inspect”. An extension of these method, system and/or apparatus is termed as“snap2inspect3D”. A flow chart illustrating the operation principles of the snap- to-inspect example is illustrated in Figure 2.
With reference to Figure 2, an image capturing unit (e.g. camera, sensor and the like) of a mobile device (such as smartphone) or an image capturing device is instructed to capture a first image 204 of a target object. For instance, such instruction is inputted by a user 202 or issued by a command received from a control unit of the mobile device. Defect inspection algorithms 208 are used to process the image 204 to detect whether the target object has a defect 210. These defect inspection algorithms are guided by different types of visual knowledge 206, which will be elaborated later. A confidence score 215 for the detection on whether the target object has the defect 210 is computed and fed back to the user 202 to assist the user 202 or the control unit to capture one or more images (or snapping one or more images using a camera) with more optimal imaging parameters (angle, distance, etc.) while performing different automatic inspection tasks (e.g. object recognition, scene reconstruction, feature detection and the like). This means that the image capturing device determines, based on the calculated confidence score 215, whether to generate better view recommendations 220 that may include improved imaging parameters with respect to translation and/or rotation movement of the imaging capturing device. The improved imaging parameters may be generated, if required, after considering the calculated confidence score. In one example, upon receiving an input, the image capturing device is instructed to capture, according to the improved imaging parameters, a second image of the object that results in calculation of an improved confidence score. With collation of confidence scores and observation of changing trends from different images captured by the image capturing unit on a common defect on the object, better imaging observations are incrementally achieved for each specific inspection task, thereby facilitating better detection performance. This means that the collated confidence scores can be used to fine tune detection of the common defect on the object or another object (of a same type as the object). “Better imaging observations” can refer to better imaging quality achieved through application of improved imaging parameters with respect to translation and/or rotation movement of the imaging capturing device.
In another example, the calculated confidence score is generated as feedback to the user after comparing the calculated confidence score against a threshold, wherein the threshold is used to reduce a number of false detection. The improved imaging parameters can be generated after comparing the calculated confidence score against the threshold. Thereafter, upon receiving an input, the image capturing device is instructed to capture, according to the improved imaging parameters, a second image of the object that results in calculation of an improved confidence score
A computer program product (i.e. application or program) may be provided to cause the execution of the above-described defect inspection algorithm on a program controlled entity. Such computer program product and its instructions thereof may be stored in a memory accessible to a processor and be executable by the processor. The computer program product may be provided or delivered as a recording means, including a non-transitory computer readable medium, such as a memory card, USB-stick, CD-ROM, DVD or also in form of a downloadable file from a server in a network. This can, for example, be achieved through a wireless communication network by the transmission of a respective file having the computer program product. The computer program product includes a mobile application tied to an operating system of a smartphone.
In one example, a decision-theory (also known as by-contrary decision) based single-view inspection without reference images is provided where an observed geometric pattern might occur in a random image, e.g. the detection of a rare event such as a crack on the inspected surface, is based on decision theory to produce a confidence score that represents defect pattern strength (e.g. crack strength). In this example, the inspection may exploit generic knowledge that is learned beforehand in the decision-theory based single-view inspection method.“Generic knowledge” used herein refers to factual information or concepts in computer visioning system for detection of defect(s) in an object.
In another example, a classification-based single-view inspection with training images is provided where visual features can be used to classify image patches into normal or abnormal (such as non-erosion or erosion regions), based on learned knowledge from training data to produce a confidence score that represents class membership probabilities.
In yet another example, a single-view inspection with 2D reference image is provided where missing materials or geometry defects are detected by comparing the real (detected) shape against the ideal shape mapped from a reference image that has no defect to produce a confidence score that represents registration quality (e.g. number of matched feature points after registration).
In a further example, a single-view inspection with 3D reference model is provided where missing materials/geometry defects are detected by comparing the real (detected) shape against the ideal shape registered/projected from a reference 3D model that has no defect to produce a confidence score that represents registration quality (e.g. number of matched feature points after registration).
For a given visual inspection task (based on a selected mode of operation), visual features are learned based on generic knowledge or from training data regarding defects. The visioning system according to snap2inspect comprises a plurality of operation modes for selection, wherein each operation mode is for detecting one type of defect, and the processing of a captured image of an object to detect whether the object has the defect is performed according to the selected operation mode. The plurality of operation modes may include at least two of the following operation modes:
• an operation mode to detect a crack on the object using a decision-theory based algorithm (Refer to“1. Single-view inspection without training images” below for details);
• an operation mode to detect erosion on the object using a classification algorithm that trains on a plurality of training images (Refer to“2. Classification-based single-view inspection with training images” below for details);
• an operation mode to detect missing material on the object using an algorithm that matches feature points of a reference 2D image with feature points of a captured image of the object (Refer to“3. Single-view inspection with a 2D reference image” below for details); and
• an operation mode to detect missing material on the object using an algorithm that matches features points of a 3D model with feature points of a captured image of the object (Refer to “4. Single-view inspection with a 3D reference model” below for details).
Several examples of defect detection by selecting one of the plurality of operation modes are shown in Table I. Table I. Utilization of visual knowledge with respect to defect types
Figure imgf000006_0003
1. Single-view inspection without training images
In one example, there is provided a method and apparatus for single-view inspection without reference images (or training images). The method is conducted based on a“a contrario " statistical framework (also known as‘by contrary decision’). In order to observe a meaningful (or significant) geometric event/pattern, the occurrence expectation should be small in the image. This means that a geometric pattern is considered meaningful in the “a contrario " framework if expectation of its occurrence would be very small in a random image. The lower the probability of such event, the more meaningful the geometric pattern is. This means that the expected occurrence of such event (represented by e) can be set as one for a random image to reduce false detection (of defects) in the image. It should be appreciated that e can be inputted by a user or predefined by the apparatus.
Mathematical expectation is the summation or integration of a possible value from a random variable. It is also known as the product of the probability of an event occurring, denoted as P(x), and a value corresponding with an actual observed occurrence of the event. The expected value is a useful property of any random variable. The mathematical expectation is usually notated as E(X) and such expected value can be computed by overall summation of distinct values that the random variable can take. The mathematical expectation will be given by mathematical formula E(X)= å (xtpt, X P , , XnPn), where x is a random variable with a probability function, f(x), p is the probability of the occurrence, and n is the number of all possible values
Figure 3 illustrates an example of a crack on a target object (e.g. engine blade) represented by a discrete line segment c\ {p, r, a}, characterized by its starting point p (from hole), its length r e M and its direction a.
In order to be robust to shading changes on the target object that is being captured by a portable imaging device (or handheld imaging device), the intensity image is locally normalized by dividing it by its median on a small neighborhood. That is, for q— (x, y), a normalized intensity image is defined as
- Equation (1 )
Figure imgf000006_0001
'
Consequently, strength of a crack, i(cp(r, a)), is defined as the negative sum of normalized intensity of pixels on the crack of the target object. The strength of the crack is represented by mathematical equation (2) below:
Figure imgf000006_0002
Equation (2)
Let C be the set of all possible cracks in a discrete image patch /. For e > 0, a crack c is said to be e-meaningful if its number of false alarm (NFA) under a hypothesis K0 satisfies
NFA(c ) := #(? c(t(c)) < e - Equation (3)
where Fc t ) is the probability that a random variable t is larger than a given threshold. The value e corresponds to an expected number of false detections under the hypothesis 7f0 , where 7f0 is a stochastic model for unstructured data. The smaller the quantity NFA (c) is, the more meaningful the crack c. Note that a detection method is said to produce e -meaningful detections (events), when the expected number of false detections is bounded by e. See definition of #C in Equation (4) below.
To search for a candidate of crack, the starting point p is set in the w x w window (depicted as square in Figure 3) and centered at a lower vertex o on a major axis 304 of an ellipse 302 representing a hole. The w x w window represents an image patch / of a captured image (e.g. 204 of Figure 2). Details on how an image patch of a captured image can be obtained will discussed later. Oi to 05 are points on a boundary of the ellipse 302. In one example, each of these points Oi to 05 can be used as an origin to generate fan-shape scanning regions for crack detection.
The length of crack r ranges discretely from rmin to rmax : r e [rmin, rmax] . The crack direction a is within a p/2 range from the ellipse orientation: a e [f - ^, f + ^]. The number of discrete direction at a given crack length r is set to be K(r) - ^r . Therefore, the number of all possible cracks in the image patch / is
Figure imgf000007_0001
- Equation (4),
where #C is the number of all possible cracks in an image patch /, #{p e J\fw(o)} is the number of a probability that starting point p lies on the lower vertex o of the ellipse, and år=rmin K(r) is the summation of the discrete directions.
To compute the distribution of the strength t(c) of a crack, empirical distribution of /(q) is estimated from its histogram. Thereafter, the law of the strength of a crack t(cp(r, a)), i.e. sum of the r pixel contributions , is obtained by convolving r copies of the distribution m.
- Equation (5),
Figure imgf000007_0002
where * denotes the convolution operator.
In this exemplary technique, a crack is detected as e-meaningful in the image patch / if its strength t(c) is larger than the threshold, Ur, e), defined in Equation 6.
Figure imgf000007_0003
- Equation (6),
The value of e can be selected as 1 and a different threshold on t(c ) can be yielded for each value of the crack length r. Among all the e-meaningful cracks (if there are any) in the image patch /, the one with the highest strength is considered as a detected crack. An example of crack detection result is shown in Figures 4A to 4C. Figure 4A is a test image for detecting cracks according to one example of the present disclosure. On region 400 is marked out as an example. Figure 4B is an exploded view of the region 400 of Figure 4A that is marked by an inspector to indicate a crack in several apertures 402 and it matches with a detection result in Figure 4C that is indicative of whether the test image has a defect. Figure 4C shows computed crack lines 404 in the several apertures 402.
2. Classification-based single-view inspection with training images
In the following paragraphs, a classification-based single-view inspection of defects with training images is discussed. Colour, texture, and Flistogram of Gradient features are calculated from the sample images and later classified by using Support Vector Machines (SVM). A SVM is a discriminative classifier formally defined by a separating hyperplane. In other words, given labelled training data (supervised learning), the algorithm outputs an optimal hyperplane which categorizes the training data according to their labels. A SVM model is a representation of the examples (training data) as points in space, mapped so that the examples (training data) of the separate categories are divided by a clear gap that is as wide as possible. New examples (test data) are then mapped into that same space and predicted to belong to a category based on which side of the gap they fall.
Some examples of the feature vectors that can be extracted from sample image patches are listed in Table II. Table II. Features extracted from sample images patches
Feature Dimension
HSV color histogram 32
Color auto correlogram 64
Color moments 6
Gabor wavelet 48
Wavelet moments 40
Histogram of Gradient 81
Total 271
Using the SVM trained by the predetermined visual features of sample images, image patches are classified into normal/abnormal (e.g., erosion or non-erosion) regions based on learned knowledge. In this case, confidence score is given by class membership probabilities. An image patch described in the present example refers to a portion (usually a rectangle window) of a training image or a test image that is being processed for detection of a defect. The patches are automatically cropped from the original training image or the original test image with some overlaps between each patch in a sliding-window way. As illustrated with reference to Figure 6A, a plurality of image patches (e.g. 601 , 602, 605, 606 and 607) are cropped from a test image 610 such that each image patch has an overlapping portion with at least two image patches proximate to it. As an example, there is an overlap 603a between image patch 601 and image patch 602, and an overlap 604a between image patch 601 and image patch 605. Likewise, there is also an overlap 603b between image patches 602 and 606 and an overlap 603c between image patches 606 and 607. If an image patch contains a detected defect, the image patch may be highlighted (for example, displaying as a black-outlined box or a hashed-outlined box) on the original test image for viewing.
The flowchart of classification-based single-view inspection with training images is shown in Figure 5. Figure 5 indicates that training data are provided to an image capturing device (visioning system) at a step S502, and the predetermined features (such as colour, texture, and histogram of gradient) are extracted out of each training image obtained from the training data at a step S504. The extracted features of the respective training images are sent to a classifier (e.g. a support vector machine) to learn before grouping each training image accordingly at a step S506. The image capturing device is configured to output and monitor a classification result indicating whether a sample object comprises a defect and/or a number of false detections at a step S512. The classification result may be used to fine-tune the rules for classifying the training images and/or test images. Preferably, the image capturing device is configured to compute a confidence score 215 from the classification result as a feedback to the image capturing device and/or a user.
The image capturing device processes the testing data in a similar fashion as the training images. For instance, testing data are provided to the image capturing device at a step S508, and the predetermined features (such as colour, texture, and histogram of gradient) are extracted out of each captured image obtained from the testing data at a step S510. The image capturing device is also configured to compute a classification result indicating whether the test object comprises a defect at the step S506. The image capturing device is configured to output and monitor a classification result indicating whether the test object comprises a defect and/or a number of false detections at a step S512, and to fine-tune the rules to classify the training images and/or test images. Preferably, the image capturing device is configured to compute a confidence score 215 from the classification result as a feedback to the image capturing device and/or a user. An example of erosion detection result is shown in Figure 6B. Image patches indicated on a test image 610 as a plurality of black-outlined boxes in Figure 6B are erosion regions 620.
3. Single-view inspection with a 2D reference image
Geometry defects on a surface of a material, or missing portions from a surface of a material can be detected quickly from a comparison between a 2D image of the surface of the material with a single 2D reference image that has no geometry defect. This is referred to as Single-view inspection with 2D reference image in the present disclosure. In the present example pertaining to Single-view inspection with 2D reference image, the discussed technique applies to a new image and a reference image that is mono-chromatic or grayscale (instead of a full colour image).
The flowchart of single-view inspection with 2D reference image is shown in Figure 7. Firstly, a rigid transform between a reference image 702 with ideal geometry and a new image 704 is determined by registration of their corresponding key point positions (e.g. edges, corners and the like). Note that image registration is a process of transforming different sets of data into one coordinate system (space). Data may be multiple images, or images captured from different times, viewpoints or sensor. The ideal geometry 706 in the reference image 702 is transformed to a coordinate space of the real geometry 708 and missing materials/geometry defects is detected by comparing features at the corresponding positions.
A confidence score 215 is obtained based on registration quality (e.g., number of matched feature points) as a feedback. Thereafter, better observation (e.g., translation/rotation of hand-held imaging devices) recommendation is given according to the obtained scores and/or score changes over time as more images are captured for capturing a defect.
4. Single-view inspection with a 3D reference model
Geometry defects on a surface of a material, or missing portions from a surface of a material can also be detected by the vision system with a 3D reference model 802 that has no geometry defect. The flowchart of single-view inspection with 3D reference model is shown in Figure 8.
With reference to Figure 8, a new 2D image 804 is registered to the 3D model 802 by registration of their corresponding key point positions (typically represented as n). With sufficient number of corresponding points (i.e. n > 3), the optimal projective transformation, associated viewing angles and other registration parameters can be derived. This means that the registration will use optimal viewing parameters derived from the corresponded key points. Then, with the optimal registration parameters, the 3D model is projected to a (2D) projection view 806 that corresponds to the respective 2D view of the new image 804. Common techniques known to a skilled person, such as Efficient Perspective- n- Point (EPnP) method developed by Lepetit et al. in their 2008 International journal paper can be used to estimate the 2D projection view 806 that corresponds to the respective 2D view of the new image.
Missing materials and/or geometry defects can be detected by comparing the corresponding positions between a real geometry 808 detected from the new 2D image 804 and an obtained ideal geometry of the projection view 806 in a coordinate space determined by transformation.
A confidence score 215 is calculated based on registration quality (e.g., number of matched feature points). Thereafter, better observation recommendation (e.g. translation and/or rotation movement of an imaging capturing unit, a sensor, a camera or the like) is calculated and generated according to the confidence scores/score changes.
5. Experimental results
The techniques discussed above are implemented (demonstrated) in an application of defect detection of engine blades. Out of the 56 hole crack regions marked out by a human inspector, the crack detection algorithm (without reference images) found 96.4% of them. The detection rate of missing material (with 2D reference image) and corrosion/erosion (with training images) are both 100%.
Table III. Part inspection results
Figure imgf000009_0001
The techniques disclosed herein can also be used in imaging sensor planning in robotic vision. The objective of imaging sensor planning in robotic vision can be summarized as follows:
With reference to Figure 9A, if information about the environment (e.g., the object 902 under observation, the available sensors 904) as well as information about the task(s) 906 that a vision system is to accomplish (i.e., detection of certain object features, object recognition, scene reconstruction, object manipulation) is provided during sensor planning, a sensor planning system 900 is able to develop strategies to automatically determine sensor parameter values 910 that can achieve a task with a certain degree of satisfaction. With such strategies, sensor parameters values 910 can be selected and can be purposefully changed in order to effectively perform the task at hand. Sensory systems are then able to operate more flexibly, autonomously, and reliably. The process of sensor planning system is plotted in Figure. 9A.
One existing model of sensor planning that is discussed in International Patent Application No. PCT/SG2017/050175 is illustrated in Figure 9B. The model in Figure 9B comprises three parts: system setup 952 for different types of objects, task specification 954 for objects of the same type and online inspection 956 for each object. The individual steps for inspection 956 are not relevant to the present disclosure and are not elaborated herein. It should be appreciated that various sensing and inspection steps can be used for the online inspection 956 in the existing method of sensor planning.
The system setup 952 consists of the following two parts:
• Sensor calibration: Calibrate different sensors (texture sensor, geometry sensor) into a same coordinate system, i.e. the robotic base coordinate system.
• System configuration: Configure device specifications, which will be considered in a viewpoint calculation. Specification includes:
1. Geometry sensor: Field of view, clear distance, depth of field, precision for each measurement point, point density at different sensing distance.
2. Texture sensor: Field of view, Depth of field
3. Robot: Position accuracy.
The task specification 954 seeks to specify global parameters 960 for objects of the same type, which includes the desired geometry and texture quality. An object model 958 for the objects to be inspected has to be prepared. If sensors and the robot (the actuator) are fixed for inspecting different types of defects on different objects (like in examples of the present disclosure), system configuration remains unchanged, and thus, the sensors and the robot only need to be calibrated once. For inspecting objects of a same type (like in examples of the present disclosure), the task specification can be performed once. If new training data is available, prior knowledge training can be updated to include the new training data.
The differences between Snap2lnspect, which refers to techniques proposed in examples of the present disclosure, and existing sensor planning methods are summarized in Table IV below.
Table IV. Comparison between with Snap2lnspect and existing sensor planning methods.
Figure imgf000010_0001
There are two main differences of the techniques disclosed in examples in the present disclosure and the existing sensor planning methods: 1 ) The criteria for optimal sensor setup in existing sensor planning methods is geometry- oriented (which aims at optimal output geometry quality). In contrast the criterion for the techniques disclosed herein is defect-oriented which aims at better detection rate.
2) Although the vision tasks to be achieved by existing sensor planning include object recognition and feature detection (which is similar to (at least) one of the tasks implemented through the techniques disclosed herein), existing sensor planning is reliant on low level sensing parameters such as resolution during the actual planning process. In contrast, the techniques disclosed herein focus on defect-level pattern recognition tasks and the corresponding detected result is actually returned as a feedback during the imaging process.
The key features of a proposed solution based on the above exemplary techniques of examples of the present disclosure are as follows
• An automatic and efficient method to present a quick“evaluation” to the inspector (or user) when mobile imaging device is positioned quickly. The‘by contrary decision’ approach is extended to the detection of any parameterizable geometric structures. The threshold is automatically determined so that parameter setting is minimal.
• In different operation modes, the input to the visual system can be generic knowledge, training image data or reference images. These different operation modes are non-model based (i.e. no sensor model and/or object model are used), and such non-model-based approaches are used to determine the best next view and sensor settings to incrementally acquire the object information. This is different from conventional sensor planning methods, because these methods use model- based approaches that require sensor model and object model to determine the optimal sensor placements and a shortest path through these viewpoints. Defect-oriented confidence scores are used to guide the inspector (user) to find the optimal imaging parameters and it is applicable to generic visual inspection of objects, structures & architectures. There is no requirement for the sensor set-up to be optimized before detection of defects. The defect-oriented confidence scores are designed for defect-level pattern recognition tasks. However, existing sensor planning methods are geometry-oriented and have to ensure that captured images have unobstructed view, are in focus, are properly magnified, are well-centered within a field-of-view, and imaged with sufficient contrast.
The advantages of the proposed solution are as follows:
• Applicable to generic visual inspection of objects or structures;
• Different operation modes: with or without reference image/model, with or without training data, etc.; and
• Automatic and efficient.
A toolbox based on the proposed solution, is useful for industries that require defect inspection with mobile (especially hand-held) imaging devices. The applications may include:
• Application on mobile imaging device for abnormity and/or defect inspection;
• Assistance to inspector to achieve better inspection performance with interactions; and
• Adaptation to task specific inspection where users can further develop based on teachings of the present disclosure. For example, classification-based single-view inspection with training images can be adapted for detecting other surface defects, such as, discoloration and burning, by using a classifier that is trained by a new set of training images of relevant defects.
6. Snap2lnspect3D
There is also provided a method, apparatus and/or a (visual or vision) system for visual inspection with active 3D measurement. Such method, apparatus and/or a (visual or vision) system is referred herein as Snap2lnspect3D. Once defects in 2D images are detected, for instance, through the snap2inspect or snap2inspect2D techniques disclosed previously, the targeted defect regions can be measured actively with an adaptive illumination pattern. This exemplary technique provides key 3D measurements such as geometry of the inspected surface in absolute scale and profile data of defects (length, width and depth) while keeping the number of image acquisitions low or within to a few snaps. It should be appreciated that this exemplary technique provides a lower cost, and yet faster method to obtain accurate 3D measurement of a detected defect as compared to existing methods.
One limitation in 3D reconstruction is known as correspondence problem, particularly when observing non-textured objects. In order to cope with the correspondence problem, methods based on structured light can be used to create correspondences and give specific code word to every position on an image. A structured light pattern is projected onto a scene and imposes illusion of texture on an object, thereby increasing the number of correspondence on non-textured objects. Therefore, surface reconstruction is possible when looking for differences between projected and recorded patterns. Coded structured light is preferred because each element of the pattern can be unambiguously identified on the images. The aim of coded structured light is to robustly obtain a large set of correspondences per image independently of the appearance of the object being illuminated and the ambient lighting conditions.
A set of patterns may be successively projected onto a surface. In such an example, the codification is temporal. Such kind of patterns can achieve high accuracy in measurements, but the ability to measure moving surfaces is restricted to one-shot scans (i.e. not robust to motion acquisition). While classical-based multiplexing techniques like M-array based techniques perform one-shot 3-D reconstruction with good accuracy results, these techniques produce sparse (feature wise) reconstructions.
The following example demonstrates that by targeting at defect regions, it is possible to perform motion-robust and accurate 3D measurement of targeted defect using a hand-held device.
The workflow of the visual inspection with active 3D measurement, or Snap2lnspect3D is shown in Figure 10. With reference to Figure 10, first, an image is captured (S1 ) using an image capturing unit of a portable imaging device (e.g. a camera on a smartphone/tablet). Once a potential defect is identified by real-time image analysis, the surface geometry within the field of view is acquired in order to perform accurate 3D measurements. To acquire surface geometry using a portable imaging device (and preferably a hand-held imaging device), a spatial coded pattern is used to reconstruct (S2) sparse 3D points of an underlying surface of a detected defect region within a single snap (or by capturing a single image of the defect region). Based on the underlying surface geometry, the image captured by the camera can be overlaid with the surface as texture and the absolute scale of the defect on the image can be inferred. Guided by defect detection (S3), the system projects an adaptive pattern to measure the target region of the defect (i.e. a defect region) for dense and accurate 3D profiling (S4).
Flence, in the workflow, upon detection that an object has a defect, for instance, by the snap2inspect3D techniques, a light projector is instructed to project a coded structured pattern on the object. Thereafter, the image capturing device is instructed to capture a two dimensional image of the object comprising the projected coded structured pattern. A three dimensional (3D) model of the object is constructed from the captured two dimensional image to obtain surface geometry of the object to facilitate 3D measurements of the defect. Optionally, the coded structured pattern is a monochromatic spatially coded pattern.
To achieve optimized measurement over the defect region, we can adapt a coded projection pattern based on a viewpoint between defect and image capturing device, tune the colours of the pattern with regard to the one of the defect regions and increase spatial sample resolution for the targeted defect region. It provides key 3D measurements such as geometry of the inspected surface in absolute scale and the profile of defects (length, width and depth) while keeping the scanning within a few shoots. Finally, task-specified dimensional criteria (S5), such as the maximum acceptable width of cracks or the depth of a hole, can be taken into account in defect verification and decision-making. The minimum number of shoots required to perform 3D measurement is dependent on complexity of the obtained surface geometry of the target object. For example, one shoot is sufficient for a target object with spherical surfaces.
A hardware set-up of a proposed vision system 1 100 is shown in Figure 1 1 . It comprises of an image capturing device 1 105, such as in the present example, a camera of a smart phone, and a portable projector 1 1 10. To set up the vision system 1 100, the image capturing device 1 105 is mounted on an adjustable clamp (not shown in Figure 1 1 ) together with the projector 11 10 at a certain angle q between the imaging capturing device 1 105 and a projected pattern 1 1 15 based on a target object 1 120 to be measured. A cable may be used to connect between an input/output port (e.g. USB slot) on the image capturing device 1 105 and an input/output port (e.g. FIDMI slot) on the projector 1 1 10 to enable data transfer between them. For applications using coded structured light, the active device is typically a digital light projector and is modelled as the inverse of a camera. When a camera is used, light from the environment is focused on an image plane and captured. Such process reduces the dimensions of the data taken in by the camera from three dimensions to two dimensions (i.e. light from a 3D scene is stored on a 2D image), and hence a system calibration is required before dimensions of a defect on a 3D object can be measured from the camera images.
Accuracy and simplicity of such system calibration is a key challenge in building a 3D measurement system 1 100. Figure 12 shows coordinate systems to be considered for the projector 1 1 10 and the camera 1 105 for system calibration. As illustrated in the example of Figure 12, the dimensions of a three-dimensional (3D) point P (Xw, Yw, Zw) taken in by a camera 1 105 is reduced from three dimensions to two dimensions, and represented as p (uc,vc). The coordinates of the point P is computed with respect to a 3D coordinate space 1202 formed by three axes Xw ,YW , Zw , which intersect at an origin O. The three axes Xw, Yw, Zw are orthogonal to one another. The coordinates of p (uc ,vc) are formed with respect to a 2D camera coordinate space formed by two axes u and v, which intersect at an origin o. The origin o of the 2D camera coordinate space corresponds to an origin Oc of a camera coordinate frame formed by three axes Xc, Yc, Zc. The three axes Xc, Yc, Zc are orthogonal to one another. The 3D point P has a corresponding projecting point p’ (up, vp) at the projector 1 1 10. The coordinates of p’ (up, vp) are formed with respect to a 2D projector coordinate space formed by two axes u and v, which intersect at an origin o. The origin o of the 2D projector coordinate space corresponds to an origin Op of a projector coordinate frame formed by three axes Xp, Yp, Zp. The three axes Xp, Yp, Zp are orthogonal to one another.
During calibration of the 3D measurement system 1 100 of Figure 1 1 as shown in Figure 12, intrinsic parameters of the camera 1 105 and the projector model 1 1 10, and extrinsic parameters between the projector 1 110 and the camera 1 105 are determined. The terms“camera” and“camera model” and the terms“projector” and “projector model” are used interchangeably in the present example. The intrinsic parameters of the camera model 1 105 include a focal length and co-ordinates of a principal point, which describes a relationship between a 3-D point P (Xw, Yw, Zw) and a corresponding projection onto an image plane. The extrinsic parameters include rotation and translation between the coordinate systems of the camera and the projector, which describes the relative position and orientation between them. In the example of Figure 12, calibration can be performed by using a black image with an illuminated pixel as a pattern. In the example, only one point can be reconstructed through triangulation using pixel coordinates of the illuminated pixel in the pattern and the corresponding coordinates in the camera image.
The intrinsic parameters of the camera model 1 105 can be estimated by widely used calibration processes, such as Direct linear transformation (DLT) method, Zhang’s method and the like. For the extrinsic calibration, the system may use a method of projecting a calibration pattern such as through the use of a planar checkerboard to minimize calibration complexity and cost under a fixed setting with the camera 1 105 and the projector 11 10 fixed. An example of the planar checkerboard is shown in Figure 13d. In another example, in order to adapt the calibration to a mobile setting such as a hand-held image capturing device, the planar checkerboard as shown in Figure 13d can be replaced with asymmetrical circle patterns as shown in Figure 13e for the ease of real-time detection.
The proposed system 1 100 and the corresponding method discussed above consider the projector as an inverse camera which maps 2D image intensities into 3D rays, and thus make the calibration of a projector-camera system a similar procedure as that of stereo cameras. In this way, by having the 2D projected points and its 3D correspondences, the system 1 100 can be calibrated using any standard stereo camera calibration method.
With regard to light coding, for motion-robust acquisition of surface geometry under a mobile setup, a monochromatic spatially coded pattern is preferred. Colored patterns can also be used although colored patterns are generally less robust than monochromatic patterns. For instance, in one example, the coded pattern can be adapted based on an M-array based theory, with uniqueness of the code for each 3 by 3 window. An algorithm such as that is proposed by Albitar et al; IEEE int. conf. on image processing, pages 529-532, 2007 (hereinafter“Albitar et al”) may be used. An alphabet of three symbols that is associated to geometrical shapes (disc, circle, and stripe) may be used. The length of the code associated to each primitive is 9 since a 3x3 window centred on this primitive is considered in this example. In one example, an obtained matrix of dimensions 27x29, which means 783 primitives, verifying first three desired criteria and using three symbols is obtained through the algorithm applied (see Figure 13a). The first three desired criteria refers to the three desired criteria used in the algorithm proposed by Albitar et al to ensure that the system decodes the coded patterns correctly are central symmetry, uniqueness of the code of each 3x3 window, and hamming distance being more than 3 (i.e. each code word is different from each other by at least three symbols). Projected pattern needs to be segmented and decoded for 3D reconstruction (see Figures 13b and 13c). First, the contours are segmented and then classified into three pattern primitives. Once each primitive together with its neighbors is detected, its code is determined. Finally, the 3D position of each decoded point can be calculated through triangulation between the observed pattern and the original one.
According to the light coding description above, the projector 1 1 10 can be configured to project a structured coded pattern on a 2D image captured by the camera 1105 for defect detection. Thereafter, a 3D model is reconstructed by decoding the projected coded pattern as discussed to obtain 3D coordinates to detect a defect on the 2D image. Once a defect is identified on the 2D image with its underlying surface parameters estimated using the method described above, the 3D coordinates of any points in the defect region can be calculated from the 2D image.
Thereafter, the system 1 100 is configured to generate a defect-specified pattern to measure the defect automatically. Figure 14 shows essentially the same elements as Figure 12 but in addition, there is shown a cube 1404 provided for illustrative purposes and it represents the 3D object to be scanned for defects. As shown in Figure 14, the system 1 100 is configured to assume a point, p, as a centre of defect in a 2D image, with its 3D point, P (XW,YW,ZW), in a camera frame 1402 (i.e. video frame captured by the camera 1 105). Based on the camera-projector calibration,
p' = K-¾ T P, - Equation (7),
where p’ is the corresponding point for the projector, K is the intrinsic parameters for the projector and T is the extrinsic parameter between camera frame 1402 and projector frame 1406.
Various pattern codification strategies can be applied. For example, with reference to Figure 17, to measure a depth of a dent on a surface, a pattern of four-points are required, where three points form a reference plane with the remaining one point as target point. Thus, the depth of the dent can be calculated as distance between the target point and the reference plane.
In order to obtain accurate 3D measurements, system calibration should be performed. As discussed with reference to Figures 1 1 and 12, to perform the calibration between the camera 1 105 and projector 1 1 10, a user (investigator) may point both the camera 1 105 and projector 1 1 10 toward a calibration board with asymmetrical circle patterns (e.g. as shown in Figure 13e). The detected patterns are automatically marked on a screen 1502 by the system 1 100 is illustrated in Figure 15, and recorded for correlation. The calibration is repeated at least a further two times, with the camera 1 105 and the projector (of the system 1 100) 1 1 10 tilted at a different angle Q toward the calibration board with asymmetrical circle patterns to collect at least three sets of data. The system calibration result with 10 sets of data is illustrated in Figure 16.
In other words, with reference to Figure 15, generally, the image capturing device 1 105 and the light projector 1 1 10 are directed towards a calibration board comprising an area 1510 with calibration pattern 1506 such that the light projector 1 1 10 projects asymmetrical circle patterns 1504 outside the area 1510 with calibration pattern 1506. The calibration pattern 1506 and the projected asymmetrical circle patterns 1504 are detected by capturing, using the image capturing device 1 105, an image of the calibration board comprising the calibration pattern 1506 and the projected asymmetrical circle patterns 1504 at a first instance. The detection of the calibration pattern 1506 and the projected asymmetrical circle patterns 1504 is graphically represented as lines 1512 and 1508 respectively, In this example, the line 1508 is formed by joining each center point of each asymmetrical circle of the pattern 1504 together, whereas the line 1512 is formed by joining each center point of each asymmetrical circle of the pattern 1506 together. Thereafter, the image capturing device 1 105 and the light projector 1 1 10 are positioned or tilted at a different angle q towards the calibration board for at least two more times. It should be appreciated that when the image capturing device 1 105 and the light projector 1 1 10 are positioned or tilted differently, the calibration pattern 1506 and the projected asymmetrical circle patterns 1504 are detected again for each additional image captured at the different angle q by the image capturing device 1 105. After capturing at least three images at respective different angles, a calibration result is obtained based on collated data of the detected calibration pattern and detected projected asymmetrical circle patterns. The calibration pattern 1506 may be different or the same as the projected asymmetrical circle patterns 1504. For example, the calibration pattern 1506 can be asymmetrical circle patterns or planar checkerboard pattern. In another example, the calibration pattern 1506 in the area 1510 may be overlaid with the structured coded pattern projected by the light projector 1 110 during calibration.
Following the completion of the calibration procedure (and taking into consideration the calibration result), the system 1 100 can perform dimensional measurement using stereo-triangulation on the selected object. For example, a height of an object on a surface, or a depth of a cavity in a surface can be measured. Figure 17 illustrates an example of height measurement of a rectangular object 1704 on a surface 1706. Four illuminated dots (or points) 1702 are projected onto the object 1704 placed on the surface 1706, and their 3D coordinates 1708 with respect to a centre of the camera 1 105 is calculated in real-time based on the triangulation. Height or depth on the upper left corner, for example, (Xi , y^ z ), is calculated by the distance from a centre point 1710 to the plane formed by the rest of three points, for example, (x0, y0, z0), (x2, y2, z2), and (x3, y3, z3). In other words, generally, the four illuminated dots 1702 are projected onto a surface containing an elevated or a dented point for measuring respective height or depth of the point relative to the surface, wherein three of the four illuminated dots 1702 for forming a reference plane are projected on the surface and the remaining one illuminated dot (e.g. the centre dot or point 1710 in Figure 17) is projected on the point. After obtaining 3D coordinates of each of the four illuminated dots 1702 with respect to a centre of the image capturing device 1 105 through triangulation, the respective height or depth of the point relative to the surface based on distance between the point and the reference plane can be calculated.
In the aforementioned examples relating to a toolbox (an application that may include a mobile application of a smartphone) for performing single-view mobile inspection of defects which is guided by different types of visual knowledge (Snap2lnspect), a confidence score is provided to assist an inspector to capture images using optimal imaging parameters (angle, distance, etc.) when performing different automatic inspection tasks. With the scores and their changing trends, better imaging observations are incrementally achieved for the specific inspection task, facilitating better detection performance.
Monocular vision might not be able to resolve the scale of the object. This presents a challenge while performing visual inspection tasks at various times, because detailed defect analysis may require measurement in absolute scale. The proposed method and the corresponding system 1 100 resolves the scale of the inspected object by stereo-triangulation with an additional projector. A flow chart 1800 in Figure 18 describes how the proposed method performed through the corresponding system 1 100 fits into the framework of Snap2lnspect3D.
Specifically, with reference to Figure 18, an inspector 1802 obtains an image of a target object and performs a one shot-scan for 3D surface reconstruction. Thereafter, Snap2lnspect3D technique 1804 described above is applied to detect a defect of the target object using visual knowledge and surface geometry involving 3D measurement (1806). That is, Snap2lnspect3D 1804 involves active sensing for surface geometry and 3D defect measurement (1808). In one example, four points may be illuminated on a detected defect region, and a 3D measurement of a defect can be subsequently made if positions of the projector 1 1 10 of Figure 1 1 and camera 1 105 of Figure 1 1 are fixed. A confidence score 1810 for the detection on whether the object has a defect is computed and fed back to an inspector 1802 to assist the inspector 1802 to capture one or more images with more optimal imaging parameters (angle, distance, etc.). After the confidence score 1810 is computed, documentation of the defect finding takes place (1812). The system 1 100 then determines, based on the calculated confidence score 1810, whether to generate better view recommendations 1814. For example, better view recommendations 1814 may include improved imaging parameters with respect to translation and/or rotation movement of the camera 1 105 and generates the improved imaging parameters if required. Upon receiving an input, the camera 1 105 may be instructed to apply the improved imaging parameters to capture a second image of the object that results in calculation of an improved confidence score. With the collated confidence scores and the observed changing trends, better imaging observations can be incrementally achieved for each specific inspection task, thereby facilitating better detection performance.
Traditional precision industries measure 3D profile of a surface using high-end 3D scanners, which require to be mounted on a flexible holder, or a hand-held scanner, which tracks the position of the scanner based on the texture on the object surface. Flowever, acquisition of accurate 3D surface profile takes a long scan time with those high-cost 3D scanners, and additional efforts are required to extract and analyze the regions of 3D surface profile related to defects by manually drawing boundary of regions or reference points. To address this problem, a 3D surface inspection tool was introduced. Such 3D surface inspection tool can scan the surface and analyze three defined types of defects: dents, rivet flushness-levels and gaps. The detection result is subsequently displayed on the measured surfaces. Flowever, such a system is designed to measure specifically the above- mentioned three types of defects. The conventional system is also bulky because a pole is required to fix the inspection system against the surface during scanning, and cannot be hand held.
The proposed method and corresponding system 1 100 of Figure 1 1 for visual inspection with active 3D measurement offers active defect measurement guided by defect detection, which could be readily implemented using existing hardware, such as smartphone or tablet with image capturing capabilities, and a hand-held projector into a low-cost palm-sized platform for on-site visual inspection. The proposed method and corresponding system 1 100 further allows active 3D measurement with adaptive illumination pattern guided by the image-based defect detection. It provides key 3D measurements such as geometry of the inspected surface in absolute scale and profile of defects (length, width and depth) while keeping the scanning time within a few snaps.
The proposed method and corresponding system 1 100 advantageously provides intelligent defect detection with absolute scale, where the image captured by the camera 1 105 can be overlaid with the surface geometry as texture. Based on the underlying geometry, the absolute scale of the detected area on the image can be inferred. Thus, task-specified dimensional criteria such as the maximum acceptable width of crack can be taken into consideration during defect detection and decision making process. The image capturing device 1 105 in Figure 1 1 may be a computing or mobile device, for example, smart phones, tablet devices, and other handheld devices. The image capturing device 1 105 in Figure 1 1 and the projector 1 1 10 in Figure 1 1 may be able to communicate through other communications network, such as, wired network, but these are omitted from Figure 1 1 for the sake of clarity. There may be provided software, such as one or more computer programs being executed within the image capturing device 1 105, and instructing the image capturing device 1 105 to connect and communicate with the projector 1 1 10 according to the operations described with reference to the earlier figures. Likewise, there may be provided software, such as one or more computer programs being executed within the projector 1 1 10, and instructing the projector 1 1 10 to connect and communicate with the image capturing device 1 105 according to the operations described with reference to the earlier figures.
The architecture of a system or an apparatus proposed in an example of the present disclosure may be an apparatus 1902 in Figure 19 and comprises a number of individual components including, but not limited to, processing unit 1916 (or processor), a memory 1918 (e.g. a volatile memory such as a Random Access Memory (RAM) for the loading of executable instructions 1920, the executable instructions defining the functionality the apparatus 1902 carries out under control of the processing unit 1916. The apparatus 1902 also comprises a network module 1925 allowing the apparatus 1902 to communicate over the communications network 1908 (for example the internet, Bluetooth network). User interface 1924 is provided for user interaction and may comprise, for example, conventional computing peripheral devices such as display monitors, computer keyboards and the like. The apparatus 1902 may also comprise a database 1926 to store defect detection results (and corresponding confidence scores and/or system settings of the image capturing device 1905 as well as the projector 1 1 10). The database 1926 may also be configured to store a coded projection pattern for projecting on a defect region of an object. It should also be appreciated that the database 1926 may not be local to the apparatus 1902. The database 1926 may be a cloud database.
The processing unit 1916 is connected to input/output devices such as a computer mouse, keyboard/keypad, a display, headphones or microphones, a video camera and the like (not illustrated in Figure) via Input/Output (I/O) interfaces 1922. The components of the processing unit 1916 typically communicate via an interconnected bus (not illustrated in Figure 19) and in a manner known to the person skilled in the relevant art.
The processing unit 1916 may be connected to the network 1908, for instance, the Internet, via a suitable transceiver device (i.e. a network interface) or a suitable wireless transceiver, to enable access to e.g. the Internet or other network systems such as a wired Local Area Network (LAN) or Wide Area Network (WAN). The processing unit 1916 of the apparatus 1902 may also be connected to one or more external wireless communication enabled device 1904 through the respective communication links 1910 and 1912 via the suitable wireless transceiver device e.g. a WiFi transceiver, Bluetooth module, Mobile telecommunication transceiver suitable for Global System for Mobile Communication (GSM), 3G, 3.5G, 4G telecommunication systems, or the like.
In another example, the architecture of a system or an apparatus proposed in an example of the present disclosure may be a device 1904 in Figure 19. The device 1904 may comprise a number of individual components including, but not limited to, microprocessor 1928 (or processor), a memory 1930 (e.g. a volatile memory such as a RAM) for the loading of executable instructions 1932, the executable instructions defining the functionality the device 1904 carries out under control of the microprocessor 1928. The device 1904 also comprises a network module (not illustrated in Figure) allowing the device 1904 to communicate over the communications network 1908. User interface 1936 is provided for user interaction and control that may be in the form of a touch panel display and presence of a keypad as is prevalent in many smart phone and other handheld devices. The device 1904 may also comprise a database (not illustrated in Figure), which may not be local to the device 1904 but a cloud database. The device 1904 may include a number of other Input/Output (I/O) interfaces 1934 as well but they may be for connection with headphones or microphones or projectors, Subscriber identity module (SIM) card, flash memory card, USB based device, and the like, which are more for mobile device usage.
The software and one or more computer programs stored may include, for example, applications that may include one or more applications for e.g. internet accessibility, operating the device 1904 and apparatus 1902 (i.e. operating system), network security, file accessibility, database management, which are applications typically equipped on a desktop or portable (mobile) device. The software and one or more computer programs may be supplied to a user of the device 1904 and/or the apparatus 1902 encoded on a data storage medium such as a CD-ROM, on a flash memory carrier or a Flard Disk Drive, and are to be read using a corresponding data storage medium drive for instance, a data storage device (not illustrated in Figure 19). Such application programs may also be downloaded from the network 1908. The application programs are read and controlled in its execution by the processing unit 1916 or microprocessor 1928. Intermediate storage of program data may be accomplished using RAM 1920 or 1930.
Furthermore, one or more of the steps of the computer programs or software may be performed in parallel rather than sequentially. One or more of the computer programs may be stored on any machine or computer readable medium that may be non-transitory in nature. The computer readable medium may include storage devices such as magnetic or optical disks, memory chips, or other storage devices suitable for interfacing with a general purpose computer or mobile device. The machine or computer readable medium may also include a hard-wired medium such as exemplified in the Internet system, or wireless medium such as exemplified in the Wireless LAN (WLAN) system. The computer program when loaded and executed on such a general-purpose computer effectively results in an apparatus (e.g. 1 105 and/or 1 1 10 in the respective Figures) that implements the steps of the computing methods in examples herein described.
In summary, examples of the present disclosure may have the following features.
A method for detecting a defect on an object, the method comprising:
instructing an image capturing device to capture a first image of an object;
processing the captured first image to detect whether the object has a defect;
calculating a confidence score for the detection on whether the object has the defect;
generating the calculated confidence score as a feedback to a user; and
determining based on the calculated confidence score whether to generate improved imaging parameters with respect to translation and/or rotation movement of the image capturing device;
generating the improved imaging parameters if required after considering the calculated confidence score; and
receiving an input to instruct the image capturing device to capture, according to the improved imaging parameters, a second image of the object that results in calculation of an improved confidence score.
The method may further comprise:
collating confidence scores obtained from different images captured by the image capturing device for a common defect on the object; and
using the collated confidence scores to fine tune detection of the common defect on the object or another object.
The method may further comprise:
selecting an operation mode from a plurality of operation modes so that the processing of the captured first image to detect whether the object has the defect is performed according to the selected operation mode,
wherein each operation mode is for detecting one type of defect, wherein the plurality of operation modes includes at least two of the following operation modes:
an operation mode to detect a crack on the object using a decision-theory based algorithm; an operation mode to detect erosion on the object using a classification algorithm that trains on a plurality of training images;
an operation mode to detect missing material on the object using an algorithm that matches features points of a reference two dimensional (2D) image with feature points of a captured image of the object; and
an operation mode to detect missing material on the object using an algorithm that matches features points of a three dimensional (3D) model with feature points of a captured image of the object.
The calculated confidence score may generated as feedback to the user after comparing the calculated confidence score against a threshold, wherein the threshold is used to reduce a number of false detection.
The method may further comprise:
upon the detection that the object has the defect, instructing a light projector to project a coded structured pattern on the object;
instructing the image capturing device to capture a two dimensional image of the object comprising the projected coded structured pattern;
constructing a three dimensional (3D) model of the object from the captured two dimensional image by decoding the projected coded structured pattern to obtain surface geometry of the object; and
performing 3D measurements of the defect using the surface geometry obtained.
The coded structured pattern may be a monochromatic spatially coded pattern.
The method may further comprise: adapting the coded structured pattern based on viewpoint between a defect and the image capturing device; and/or
tuning colours of the coded structured pattern in regard to a specific defect region; and/or increasing spatial sample resolution for a target defect region considered.
The method may further comprise:
directing the image capturing device and the light projector towards a calibration board comprising an area with calibration pattern such that the light projector projects asymmetrical circle patterns outside the area with calibration pattern;
detecting the calibration pattern and the projected asymmetrical circle patterns by capturing using the image capturing device an image of the calibration board comprising the calibration pattern and the projected asymmetrical circle patterns;
tilting the image capturing device and the light projector at a different angle towards the calibration board for at least two more times;
detecting the calibration pattern and the projected asymmetrical circle patterns for additional images captured at the different angles by the image capturing device; and
obtaining a calibration result based on collated data of the detected calibration pattern and detected projected asymmetrical circle patterns.
The calibration pattern may be a plurality of asymmetrical circle patterns.
The method may further comprise:
taking into consideration the calibration result, performing measurement using stereo triangulation on an image of the object captured by the image capturing device.
The method may further comprise:
projecting four illuminated dots onto a surface containing an elevated or a dented point for measuring respective height or depth of the point relative to the surface, wherein three of the four illuminated dots for forming a reference plane are projected on the surface and the remaining one illuminated dot is projected on the point;
calculating 3D coordinates of the four illuminated dots with respect to a centre of the image capturing device based on triangulation; and
after obtaining the 3D coordinates of the four illuminated dots, calculating height or depth of the point relative to the surface based on distance between the point and the reference plane.
An apparatus for detecting a defect on an object, the apparatus comprises:
a processor configured to execute instructions in a memory to control the apparatus to:
instruct an image capturing device (e.g.1 105 of Figure 1 1 ) to capture a first image (e.g. 204 of Figure 2) of an object (e.g. 1704 of Figure 17);
process the captured first image to detect whether the object has a defect (e.g. 210 of Figure
2) ;
calculate a confidence score (e.g. 215 of Figure 2) for the detection on whether the object has the defect;
generate the calculated confidence score as a feedback to a user (e.g. 202 of Figure 2); determine based on the calculated confidence score whether to generate improved imaging parameters (e.g. 220 of Figure 2) with respect to translation and/or rotation movement of the image capturing device;
generate the improved imaging parameters if required after considering the calculated confidence score; and
receive an input to instruct the image capturing device to capture, according to the improved imaging parameters, a second image of the object that results in calculation of an improved confidence score.
The apparatus may be controllable to:
collate confidence scores obtained from different images captured by the image capturing device for a common defect on the object; and
use the collated confidence scores to fine tune detection of the common defect on the object or another object.
The apparatus may be controllable to select an operation mode from a plurality of operation modes so that the processing of the captured first image to detect whether the object has the defect is performed according to the selected operation mode,
wherein each operation mode is for detecting one type of defect, wherein the plurality of operation modes includes at least two of the following operation modes:
an operation mode to detect a crack on the object using a decision-theory based algorithm; an operation mode to detect erosion on the object using a classification algorithm that trains on a plurality of training images an operation mode to detect missing material on the object using an algorithm that matches features points of a reference two dimensional (2D) image with feature points of a captured image of the object; and
an operation mode to detect missing material on the object using an algorithm that matches features points of a three dimensional (3D) model with feature points of a captured image of the object.
The calculated confidence score may be generated as feedback to the user after comparing the calculated confidence score against a threshold, wherein the threshold is used to reduce a number of false detection.
The apparatus may be controllable to:
upon the detection that the object has the defect, instruct a light projector (e.g. 1 1 10 of Figure 1 1 ) to project a coded structured pattern on the object;
instruct the image capturing device to capture a two dimensional image of the object comprising the projected coded structured pattern;
construct a three dimensional (3D) model of the object from the captured two dimensional image by decoding the projected coded structured pattern to obtain surface geometry of the object; and
perform 3D measurements of the defect using the surface geometry obtained.
The coded structured pattern may be a monochromatic spatially coded pattern (e.g. Figure
13 A).
The apparatus may be controllable to:
adapt the coded structured pattern based on viewpoint between a defect and the image capturing device; and/or
tune colours of the coded structured pattern in regard to a specific defect region; and/or increase spatial sample resolution for a target defect region considered.
After directing the image capturing device and the light projector towards a calibration board comprising an area (e.g. 1510 of Figure 15) with calibration pattern (e.g. 1506 of Figure 15) such that the light projector projects asymmetrical circle patterns (e.g. 1504 of Figure 15) outside the area with calibration pattern, the apparatus may be controllable to:
detect the calibration pattern and the projected asymmetrical circle patterns from an image captured using the image capturing device, wherein the image comprises the calibration board comprising the calibration pattern and the projected asymmetrical circle patterns; and
after tilting the image capturing device and the light projector at a different angle towards the calibration board for at least two more times, the apparatus may be controllable to:
detect the calibration pattern and the projected asymmetrical circle patterns for additional images captured at the different angles by the image capturing device; and
obtain a calibration result based on collated data of the detected calibration pattern and detected projected asymmetrical circle patterns.
The calibration pattern may be a plurality of asymmetrical circle patterns.
The apparatus may be controllable to:
take into consideration the calibration result to perform measurement using stereo triangulation on an image of the object that is captured by the image capturing device.
After projecting four illuminated dots (e.g. 1702 of Figure 17) onto a surface (e.g. 1706 of Figure 17) containing an elevated or a dented point (e.g. 1710 of Figure 17) for measuring respective height or depth of the point relative to the surface, wherein three of the four illuminated dots for forming a reference plane are projected on the surface and the remaining one illuminated dot is projected on the point, the apparatus may be controllable to:
calculate 3D coordinates (e.g. 1708 of Figure 17) of the four illuminated dots with respect to a centre of the image capturing device based on triangulation from an image captured by the image capturing device, wherein the image comprises the four illuminated dots, the surface and the point; and
after obtaining the 3D coordinates of the four illuminated dots, the apparatus is controllable to: calculate height or depth of the point relative to the surface based on distance between the point and the reference plane.
Throughout this specification and claims which follow, unless the context requires otherwise, the word“comprise”, and variations such as“comprises” or“comprising”, will be understood to imply the inclusion of a stated integer or group of integers or steps but not the exclusion of any other integer or group of integers. While the invention has been described in the present disclosure in connection with a number of embodiments and implementations, the invention is not so limited but covers various obvious modifications and equivalent arrangements, which fall within the purview of the appended claims. Although features of the invention are expressed in certain combinations among the claims, it is contemplated that these features can be arranged in any combination and order.

Claims

Claims
1. A method for visual inspection, the method comprising:
instructing an image capturing device to capture a first image of an object;
processing the captured first image to detect whether the object has a defect;
calculating a confidence score for the detection on whether the object has the defect;
generating the calculated confidence score as a feedback to a user; and
determining based on the calculated confidence score whether to generate improved imaging parameters with respect to translation and/or rotation movement of the image capturing device;
generating the improved imaging parameters if required after considering the calculated confidence score; and
receiving an input to instruct the image capturing device to capture, according to the improved imaging parameters, a second image of the object that results in calculation of an improved confidence score.
2. The method as claimed in claim 1 , wherein the method further comprises:
collating confidence scores obtained from different images captured by the image capturing device for a common defect on the object; and
using the collated confidence scores to fine tune detection of the common defect on the object or another object.
3. The method as claimed in claim 1 , where the method further comprises:
selecting an operation mode from a plurality of operation modes so that the processing of the captured first image to detect whether the object has the defect is performed according to the selected operation mode,
wherein each operation mode is for detecting one type of defect, wherein the plurality of operation modes includes at least two of the following operation modes:
an operation mode to detect a crack on the object using a decision-theory based algorithm; an operation mode to detect erosion on the object using a classification algorithm that trains on a plurality of training images;
an operation mode to detect missing material on the object using an algorithm that matches features points of a reference two dimensional (2D) image with feature points of a captured image of the object; and
an operation mode to detect missing material on the object using an algorithm that matches features points of a three dimensional (3D) model with feature points of a captured image of the object.
4. The method as claimed in claim 1 , wherein the calculated confidence score is generated as feedback to the user after comparing the calculated confidence score against a threshold, wherein the threshold is used to reduce a number of false detection.
5. The method as claimed in claim 1 , wherein the method further comprises:
upon the detection that the object has the defect, instructing a light projector to project a coded structured pattern on the object;
instructing the image capturing device to capture a two dimensional image of the object comprising the projected coded structured pattern;
constructing a three dimensional (3D) model of the object from the captured two dimensional image by decoding the projected coded structured pattern to obtain surface geometry of the object; and
performing 3D measurements of the defect using the surface geometry obtained.
6. The method as claimed in claim 5, wherein the coded structured pattern is a monochromatic spatially coded pattern.
7. The method as claimed in claim 5, wherein the method further comprises:
adapting the coded structured pattern based on viewpoint between a defect and the image capturing device; and/or
tuning colours of the coded structured pattern in regard to a specific defect region; and/or increasing spatial sample resolution for a target defect region considered.
8. The method as claimed in claim 5, wherein the method further comprises:
directing the image capturing device and the light projector towards a calibration board comprising an area with calibration pattern such that the light projector projects asymmetrical circle patterns outside the area with calibration pattern;
detecting the calibration pattern and the projected asymmetrical circle patterns by capturing using the image capturing device an image of the calibration board comprising the calibration pattern and the projected asymmetrical circle patterns;
tilting the image capturing device and the light projector at a different angle towards the calibration board for at least two more times;
detecting the calibration pattern and the projected asymmetrical circle patterns for additional images captured at the different angles by the image capturing device; and
obtaining a calibration result based on collated data of the detected calibration pattern and detected projected asymmetrical circle patterns.
9. The method as claimed in claim 8, wherein the calibration pattern is a plurality of asymmetrical circle patterns.
10. The method as claimed in claim 8, wherein the method further comprises:
taking into consideration the calibration result, performing measurement using stereo triangulation on an image of the object captured by the image capturing device.
1 1 . The method as claimed in claim 1 , wherein the method further comprises:
projecting four illuminated dots onto a surface containing an elevated or a dented point for measuring respective height or depth of the point relative to the surface, wherein three of the four illuminated dots for forming a reference plane are projected on the surface and the remaining one illuminated dot is projected on the point;
calculating 3D coordinates of the four illuminated dots with respect to a centre of the image capturing device based on triangulation; and
after obtaining the 3D coordinates of the four illuminated dots, calculating height or depth of the point relative to the surface based on distance between the point and the reference plane.
12. An apparatus for visual inspection, the apparatus comprises:
a processor configured to execute instructions in a memory to control the apparatus to:
instruct an image capturing device to capture a first image of an object;
process the captured first image to detect whether the object has a defect;
calculate a confidence score for the detection on whether the object has the defect;
generate the calculated confidence score as a feedback to a user;
determine based on the calculated confidence score whether to generate improved imaging parameters with respect to translation and/or rotation movement of the image capturing device;
generate the improved imaging parameters if required after considering the calculated confidence score; and
receive an input to instruct the image capturing device to capture, according to the improved imaging parameters, a second image of the object that results in calculation of an improved confidence score.
13. The apparatus as claimed in claim 12, wherein the apparatus is controllable to:
collate confidence scores obtained from different images captured by the image capturing device for a common defect on the object; and
use the collated confidence scores to fine tune detection of the common defect on the object or another object.
14. The apparatus as claimed in claim 12, where the apparatus is controllable to:
select an operation mode from a plurality of operation modes so that the processing of the captured first image to detect whether the object has the defect is performed according to the selected operation mode,
wherein each operation mode is for detecting one type of defect, wherein the plurality of operation modes includes at least two of the following operation modes:
an operation mode to detect a crack on the object using a decision-theory based algorithm; an operation mode to detect erosion on the object using a classification algorithm that trains on a plurality of training images; an operation mode to detect missing material on the object using an algorithm that matches features points of a reference two dimensional (2D) image with feature points of a captured image of the object; and
an operation mode to detect missing material on the object using an algorithm that matches features points of a three dimensional (3D) model with feature points of a captured image of the object.
15. The apparatus as claimed in claim 12, wherein the calculated confidence score is generated as feedback to the user after comparing the calculated confidence score against a threshold, wherein the threshold is used to reduce a number of false detection.
16. The apparatus as claimed in claim 12, wherein the apparatus is controllable to:
upon the detection that the object has the defect, instruct a light projector to project a coded structured pattern on the object;
instruct the image capturing device to capture a two dimensional image of the object comprising the projected coded structured pattern;
construct a three dimensional (3D) model of the object from the captured two dimensional image by decoding the projected coded structured pattern to obtain surface geometry of the object; and
perform 3D measurements of the defect using the surface geometry obtained.
17. The apparatus as claimed in claim 16, wherein the coded structured pattern is a monochromatic spatially coded pattern.
18. The apparatus as claimed in claim 16, wherein the apparatus is controllable to:
adapt the coded structured pattern based on viewpoint between a defect and the image capturing device; and/or
tune colours of the coded structured pattern in regard to a specific defect region; and/or increase spatial sample resolution for a target defect region considered.
19. The apparatus as claimed in claim 16, wherein
after directing the image capturing device and the light projector towards a calibration board comprising an area with calibration pattern such that the light projector projects asymmetrical circle patterns outside the area with calibration pattern, the apparatus is controllable to:
detect the calibration pattern and the projected asymmetrical circle patterns from an image captured using the image capturing device, wherein the image comprises the calibration board comprising the calibration pattern and the projected asymmetrical circle patterns; and
after tilting the image capturing device and the light projector at a different angle towards the calibration board for at least two more times, the apparatus is controllable to:
detect the calibration pattern and the projected asymmetrical circle patterns for additional images captured at the different angles by the image capturing device; and
obtain a calibration result based on collated data of the detected calibration pattern and detected projected asymmetrical circle patterns.
20. The apparatus as claimed in claim 19, wherein the calibration pattern is a plurality of asymmetrical circle patterns.
21 . The apparatus as claimed in claim 19, wherein the apparatus is controllable to:
take into consideration the calibration result to perform measurement using stereo triangulation on an image of the object that is captured by the image capturing device.
22. The apparatus as claimed in claim 12, wherein
after projecting four illuminated dots onto a surface containing an elevated or a dented point for measuring respective height or depth of the point relative to the surface, wherein three of the four illuminated dots for forming a reference plane are projected on the surface and the remaining one illuminated dot is projected on the point, the apparatus is controllable to:
calculate 3D coordinates of the four illuminated dots with respect to a centre of the image capturing device based on triangulation from an image captured by the image capturing device, wherein the image comprises the four illuminated dots, the surface and the point; and
after obtaining the 3D coordinates of the four illuminated dots, the apparatus is controllable to: calculate height or depth of the point relative to the surface based on distance between the point and the reference plane.
PCT/SG2019/050138 2018-03-14 2019-03-14 Method for visual inspection and apparatus thereof WO2019177539A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
SG11202008944XA SG11202008944XA (en) 2018-03-14 2019-03-14 Method for visual inspection and apparatus thereof

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
SG10201802115V 2018-03-14
SG10201802115V 2018-03-14

Publications (1)

Publication Number Publication Date
WO2019177539A1 true WO2019177539A1 (en) 2019-09-19

Family

ID=67908443

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/SG2019/050138 WO2019177539A1 (en) 2018-03-14 2019-03-14 Method for visual inspection and apparatus thereof

Country Status (2)

Country Link
SG (1) SG11202008944XA (en)
WO (1) WO2019177539A1 (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111275751A (en) * 2019-10-12 2020-06-12 浙江省北大信息技术高等研究院 Unsupervised absolute scale calculation method and system
CN111383332A (en) * 2020-03-26 2020-07-07 深圳市菲森科技有限公司 Three-dimensional scanning and reconstruction system, computer equipment and readable storage medium
WO2021126074A1 (en) * 2019-12-20 2021-06-24 Ultron Techniques Pte Ltd System and method for visual inspection
CN113538411A (en) * 2021-08-06 2021-10-22 广东电网有限责任公司 Insulator defect detection method and device
CN114103494A (en) * 2020-09-01 2022-03-01 佳能株式会社 Image processing apparatus, control method thereof, and storage medium
US20220084190A1 (en) * 2020-09-15 2022-03-17 Aisin Corporation Abnormality detection device, abnormality detection computer program product, and abnormality detection system
CN114419041A (en) * 2022-03-29 2022-04-29 武汉大学 Identification method and device for focus color
WO2022256460A1 (en) * 2021-06-01 2022-12-08 Buildingestimates.Com Limited Systems for rapid accurate complete detailing and cost estimation for building construction from 2d plans
US11625553B2 (en) 2021-06-01 2023-04-11 Buildingestimates.Com Limited Rapid and accurate modeling of a building construction structure including estimates, detailing, and take-offs using artificial intelligence
EP4276747A1 (en) * 2022-05-13 2023-11-15 RTX Corporation System and method for inspecting a component for anomalous region

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1999016010A1 (en) * 1997-09-22 1999-04-01 Intelligent Reasoning Systems, Inc. Automated visual inspection system and process for detecting and classifying defects
US20060034506A1 (en) * 2004-08-13 2006-02-16 James Mahon Machine vision analysis system and method
WO2010112894A1 (en) * 2009-04-02 2010-10-07 Roke Manor Research Limited Automated 3d article inspection
US20130034298A1 (en) * 2011-08-04 2013-02-07 University Of Southern California Image-based crack detection
US20130155474A1 (en) * 2008-01-18 2013-06-20 Mitek Systems Systems and methods for automatic image capture on a mobile device
CN103175485A (en) * 2013-02-20 2013-06-26 天津工业大学 Method for visually calibrating aircraft turbine engine blade repair robot
US20160180510A1 (en) * 2014-12-23 2016-06-23 Oliver Grau Method and system of geometric camera self-calibration quality assessment

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1999016010A1 (en) * 1997-09-22 1999-04-01 Intelligent Reasoning Systems, Inc. Automated visual inspection system and process for detecting and classifying defects
US20060034506A1 (en) * 2004-08-13 2006-02-16 James Mahon Machine vision analysis system and method
US20130155474A1 (en) * 2008-01-18 2013-06-20 Mitek Systems Systems and methods for automatic image capture on a mobile device
WO2010112894A1 (en) * 2009-04-02 2010-10-07 Roke Manor Research Limited Automated 3d article inspection
US20130034298A1 (en) * 2011-08-04 2013-02-07 University Of Southern California Image-based crack detection
CN103175485A (en) * 2013-02-20 2013-06-26 天津工业大学 Method for visually calibrating aircraft turbine engine blade repair robot
US20160180510A1 (en) * 2014-12-23 2016-06-23 Oliver Grau Method and system of geometric camera self-calibration quality assessment

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111275751A (en) * 2019-10-12 2020-06-12 浙江省北大信息技术高等研究院 Unsupervised absolute scale calculation method and system
CN111275751B (en) * 2019-10-12 2022-10-25 浙江省北大信息技术高等研究院 Unsupervised absolute scale calculation method and system
WO2021126074A1 (en) * 2019-12-20 2021-06-24 Ultron Techniques Pte Ltd System and method for visual inspection
CN111383332A (en) * 2020-03-26 2020-07-07 深圳市菲森科技有限公司 Three-dimensional scanning and reconstruction system, computer equipment and readable storage medium
CN111383332B (en) * 2020-03-26 2023-10-13 深圳市菲森科技有限公司 Three-dimensional scanning and reconstruction system, computer device and readable storage medium
CN114103494A (en) * 2020-09-01 2022-03-01 佳能株式会社 Image processing apparatus, control method thereof, and storage medium
US20220084190A1 (en) * 2020-09-15 2022-03-17 Aisin Corporation Abnormality detection device, abnormality detection computer program product, and abnormality detection system
WO2022256460A1 (en) * 2021-06-01 2022-12-08 Buildingestimates.Com Limited Systems for rapid accurate complete detailing and cost estimation for building construction from 2d plans
US11625553B2 (en) 2021-06-01 2023-04-11 Buildingestimates.Com Limited Rapid and accurate modeling of a building construction structure including estimates, detailing, and take-offs using artificial intelligence
CN113538411A (en) * 2021-08-06 2021-10-22 广东电网有限责任公司 Insulator defect detection method and device
CN114419041B (en) * 2022-03-29 2022-06-21 武汉大学 Method and device for identifying focus color
CN114419041A (en) * 2022-03-29 2022-04-29 武汉大学 Identification method and device for focus color
EP4276747A1 (en) * 2022-05-13 2023-11-15 RTX Corporation System and method for inspecting a component for anomalous region

Also Published As

Publication number Publication date
SG11202008944XA (en) 2020-10-29

Similar Documents

Publication Publication Date Title
WO2019177539A1 (en) Method for visual inspection and apparatus thereof
JP6573354B2 (en) Image processing apparatus, image processing method, and program
US10288418B2 (en) Information processing apparatus, information processing method, and storage medium
JP5812599B2 (en) Information processing method and apparatus
WO2011162388A4 (en) Point group data processing device, point group data processing system, point group data processing method, and point group data processing program
JP6352208B2 (en) 3D model processing apparatus and camera calibration system
JP6503153B2 (en) System and method for automatically selecting a 3D alignment algorithm in a vision system
GB2512460A (en) Position and orientation measuring apparatus, information processing apparatus and information processing method
CN112161619A (en) Pose detection method, three-dimensional scanning path planning method and detection system
US10997702B2 (en) Inspection apparatus, inspection method, and computer readable recording medium storing inspection program
JP7339316B2 (en) System and method for simultaneous consideration of edges and normals in image features by a vision system
JP6172432B2 (en) Subject identification device, subject identification method, and subject identification program
EP4009273A1 (en) Cloud-to-cloud comparison using artificial intelligence-based analysis
JP5794427B2 (en) Marker generation device, marker generation detection system, marker generation detection device, marker, marker generation method and program thereof
US9766708B2 (en) Locating method, locating device, depth determining method and depth determining device of operating body
US11468609B2 (en) Methods and apparatus for generating point cloud histograms
US9398208B2 (en) Imaging apparatus and imaging condition setting method and program
US20230044371A1 (en) Defect detection in a point cloud
CN113125434A (en) Image analysis system and method of controlling photographing of sample image
US11717970B2 (en) Controller, control method using controller, and control system
JP2016206909A (en) Information processor, and information processing method
GB2589178A (en) Cross-domain metric learning system and method
Uyanik et al. A method for determining 3D surface points of objects by a single camera and rotary stage
US11436754B2 (en) Position posture identification device, position posture identification method and position posture identification program
EP4044107A1 (en) Upscaling triangulation scanner images to reduce noise

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19768118

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19768118

Country of ref document: EP

Kind code of ref document: A1