US20230298327A1 - Information processing device, determination method, and information processing program - Google Patents

Information processing device, determination method, and information processing program Download PDF

Info

Publication number
US20230298327A1
US20230298327A1 US18/008,872 US202118008872A US2023298327A1 US 20230298327 A1 US20230298327 A1 US 20230298327A1 US 202118008872 A US202118008872 A US 202118008872A US 2023298327 A1 US2023298327 A1 US 2023298327A1
Authority
US
United States
Prior art keywords
defect
image
determination
inspection
determining
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US18/008,872
Inventor
Takeru KATAYAMA
Kaoru Shinoda
Masamitsu Abe
Ryota IOKA
Takahiro Wada
Joichi MURAKAMI
Hiroshi Hattori
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hitachi Zosen Corp
Original Assignee
Hitachi Zosen Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hitachi Zosen Corp filed Critical Hitachi Zosen Corp
Publication of US20230298327A1 publication Critical patent/US20230298327A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N29/00Investigating or analysing materials by the use of ultrasonic, sonic or infrasonic waves; Visualisation of the interior of objects by transmitting ultrasonic or sonic waves through the object
    • G01N29/22Details, e.g. general constructional or apparatus details
    • G01N29/225Supports, positioning or alignment in moving situation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/776Validation; Performance evaluation
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01BMEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
    • G01B17/00Measuring arrangements characterised by the use of infrasonic, sonic or ultrasonic vibrations
    • G01B17/02Measuring arrangements characterised by the use of infrasonic, sonic or ultrasonic vibrations for measuring thickness
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N29/00Investigating or analysing materials by the use of ultrasonic, sonic or infrasonic waves; Visualisation of the interior of objects by transmitting ultrasonic or sonic waves through the object
    • G01N29/04Analysing solids
    • G01N29/06Visualisation of the interior, e.g. acoustic microscopy
    • G01N29/0654Imaging
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N29/00Investigating or analysing materials by the use of ultrasonic, sonic or infrasonic waves; Visualisation of the interior of objects by transmitting ultrasonic or sonic waves through the object
    • G01N29/04Analysing solids
    • G01N29/06Visualisation of the interior, e.g. acoustic microscopy
    • G01N29/0654Imaging
    • G01N29/069Defect imaging, localisation and sizing using, e.g. time of flight diffraction [TOFD], synthetic aperture focusing technique [SAFT], Amplituden-Laufzeit-Ortskurven [ALOK] technique
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N29/00Investigating or analysing materials by the use of ultrasonic, sonic or infrasonic waves; Visualisation of the interior of objects by transmitting ultrasonic or sonic waves through the object
    • G01N29/04Analysing solids
    • G01N29/11Analysing solids by measuring attenuation of acoustic waves
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N29/00Investigating or analysing materials by the use of ultrasonic, sonic or infrasonic waves; Visualisation of the interior of objects by transmitting ultrasonic or sonic waves through the object
    • G01N29/44Processing the detected response signal, e.g. electronic circuits specially adapted therefor
    • G01N29/4409Processing the detected response signal, e.g. electronic circuits specially adapted therefor by comparison
    • G01N29/4418Processing the detected response signal, e.g. electronic circuits specially adapted therefor by comparison with a model, e.g. best-fit, regression analysis
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N29/00Investigating or analysing materials by the use of ultrasonic, sonic or infrasonic waves; Visualisation of the interior of objects by transmitting ultrasonic or sonic waves through the object
    • G01N29/44Processing the detected response signal, e.g. electronic circuits specially adapted therefor
    • G01N29/4445Classification of defects
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N29/00Investigating or analysing materials by the use of ultrasonic, sonic or infrasonic waves; Visualisation of the interior of objects by transmitting ultrasonic or sonic waves through the object
    • G01N29/44Processing the detected response signal, e.g. electronic circuits specially adapted therefor
    • G01N29/4481Neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0004Industrial image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0004Industrial image inspection
    • G06T7/0008Industrial image inspection checking presence/absence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N2291/00Indexing codes associated with group G01N29/00
    • G01N2291/02Indexing codes associated with the analysed material
    • G01N2291/025Change of phase or condition
    • G01N2291/0258Structural degradation, e.g. fatigue of composites, ageing of oils
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N2291/00Indexing codes associated with group G01N29/00
    • G01N2291/02Indexing codes associated with the analysed material
    • G01N2291/028Material parameters
    • G01N2291/02854Length, thickness
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N2291/00Indexing codes associated with group G01N29/00
    • G01N2291/02Indexing codes associated with the analysed material
    • G01N2291/028Material parameters
    • G01N2291/0289Internal structure, e.g. defects, grain size, texture
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N2291/00Indexing codes associated with group G01N29/00
    • G01N2291/26Scanned objects
    • G01N2291/263Surfaces
    • G01N2291/2636Surfaces cylindrical from inside
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N2291/00Indexing codes associated with group G01N29/00
    • G01N2291/26Scanned objects
    • G01N2291/267Welds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10132Ultrasound image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30108Industrial image inspection
    • G06T2207/30148Semiconductor; IC; Wafer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30108Industrial image inspection
    • G06T2207/30164Workpiece; Machine component

Definitions

  • the present invention relates to an information processing device and the like that determine a given determination matter (matter to be determined) in accordance with target data.
  • Patent Literature 1 indicated below discloses a technique that determines presence or absence of a defect and the type of the defect in accordance with an image of a semiconductor substrate. Specifically, the technique disclosed in Patent Literature 1 decides a final classification result in accordance with a total sum obtained by summing up values obtained by multiplying (i) respective classification results of classifiers by (ii) weights respectively determined in advance for the classifiers. With this, it is possible to expect improvement in classification accuracy as compared to that of a case in which only one classifier is used.
  • the accuracy of classification by each classifier can vary depending on the image, since the images differ from each other in various points. Therefore, the weights determined in advance may not always be optimum. Using the weights that are not optimum can affect the accuracy in final determination.
  • the following case may occur. That is, when two classifiers A and B are used, classification made by the classifier A is correct for a certain image but classification made by the classifier B is incorrect for the certain image, or vice versa for another image. In this case, if the weight on the classifier A is set to be greater than that of the classification B, a final classification result on the certain image is correct, but a final classification result on the another image is incorrect.
  • Such a problem is not limited to classification involving use of a plurality of classifiers, but is a common problem that can occur when a final determination result is derived in accordance with determination results given by a plurality of determining sections with regard to a given determination matter. Further, such a problem is not limited to determination involving use of an image, but is a common problem that can occur when determination is made in accordance with arbitrary target data.
  • An aspect of the present invention was made in view of the above-described problem, and has an object to realize an information processing device and the like capable of deriving a final determination result in appropriate consideration of determination results of determining sections in accordance with target data.
  • an information processing device in accordance with an aspect of the present invention includes: a reliability determining section configured to carry out, for each of determining sections configured to determine a given determination matter in accordance with single target data, a process of determining a reliability, which is an indicator indicating a degree of certainty of a determination result given by the determining section, in accordance with the target data; and a comprehensive determination section configured to determine the given determination matter with use of the determination results and the reliabilities determined by the reliability determining section.
  • a determination method in accordance with an aspect of the present invention is a determination method that is to be executed by one or more information processing devices, including the steps of: (i) carrying out, for each of determining sections each configured to determine a given determination matter in accordance with single target data, a process of determining a reliability, which is an indicator indicating a degree of certainty of a determination result given by the determining section, in accordance with the target data; and (ii) determining the given determination matter with use of the determination results and the reliabilities determined in the step (i).
  • FIG. 1 is a block diagram illustrating an example of a configuration of a main part of an information processing device in accordance with Embodiment 1 of the present invention.
  • FIG. 2 is a view illustrating an outline of an inspection system including the information processing device.
  • FIG. 3 is a view illustrating an outline of an inspection carried out by the information processing device.
  • FIG. 4 is a view illustrating an example of a configuration of a determining section that carries out determination with use of a generative model and an example of a method for determination, carried out by the determining section, of presence or absence of a defect.
  • FIG. 5 is a view illustrating an example in which a heat map is generated from an ultrasonic testing image and threshold processing is carried out on the heat map thus generated.
  • FIG. 6 is a view illustrating a relation between the position of a defect, an ultrasonic testing image, and a heat map.
  • FIG. 7 is a view illustrating a method for detecting a defect area.
  • FIG. 8 is a view illustrating an example of areas set for respective types of defect.
  • FIG. 9 is a view illustrating a method for integrating defects captured in a plurality of ultrasonic testing images to detect the defects as a single defect.
  • FIG. 10 is a view illustrating a method for calculating a thickness of a tube-to-tubesheet weld.
  • FIG. 11 is a view illustrating an example of output of an inspection result.
  • FIG. 12 is a view illustrating an example of a process of constructing various models used in an inspection and determining a threshold.
  • FIG. 13 is a view illustrating an example of an inspection method involving use of the information processing device.
  • FIG. 14 is a flowchart illustrating an example of a defect type determining process that determines the type of a defect with use of a type decision model.
  • FIG. 15 is a flowchart illustrating an example of a defect type determining process that determines the type of a defect in accordance with the position of a defect area.
  • FIG. 2 is a view illustrating an outline of an inspection system 100 .
  • the inspection system 100 is a system that carries out an inspection to determine, in accordance with an image of an inspection target, presence or absence of a defect in the inspection target.
  • the inspection system 100 includes an information processing device 1 and an ultrasonic testing device 7 .
  • the description in the present embodiment will discuss an example in which the inspection system 100 carries out an inspection to determine presence or absence of a defect in a tube-to-tubesheet weld of a heat exchanger.
  • the tube-to-tubesheet weld refers to a part in which a plurality of metal tubes constituting the heat exchanger are welded to a metal tubesheet that bundles the tubes.
  • the defect in the tube-to-tubesheet weld refers to a gap created inside the tube-to-tubesheet weld.
  • each of the tubes and the tubesheet may be made of a nonferrous metal such as aluminum or a resin.
  • the inspection system 100 it is also possible to carry out an inspection to determine presence or absence of a defect in a welded part (base welded part) between a tube support and a tube in boiler equipment used in a garbage incineration plant, for example.
  • the part to be inspected is not limited to the welded part, and the inspection target is not limited to the heat exchanger.
  • An inspection is carried in the following manner. As shown in FIG. 2 , a probe having a contact medium applied thereto is inserted through a tube end. Then, the probe emits an ultrasonic wave so that the ultrasonic wave is propagated from an inner wall surface side of the tube toward the tube-to-tubesheet weld, and measures an echo of the ultrasonic wave. If such a defect as a gap in the tube-to-tubesheet weld occurs, an echo from the gap can be measured. In accordance with the echo, it is possible to detect the defect.
  • the lower left part of FIG. 2 shows an enlarged view of an area around the probe.
  • an ultrasonic wave indicated by the arrow L 3 is propagated in a portion of the tube-to-tubesheet weld which portion has no gap.
  • an echo of the ultrasonic wave indicated by the arrow L 3 would not be measured.
  • an ultrasonic wave indicated by the arrow L 2 is propagated toward a portion of the tube-to-tubesheet weld which portion has a gap.
  • an echo of the ultrasonic wave reflected by the gap is measured.
  • an ultrasonic wave is reflected also by the periphery of the tube-to-tubesheet weld, and therefore an echo of the ultrasonic wave propagated in the periphery is also measured.
  • an ultrasonic wave indicated by the arrow L 1 is propagated in a part closer to the tube end than the tube-to-tubesheet weld is, the ultrasonic wave does not hit the tube-to-tubesheet weld but is reflected by a tube surface of the part closer to the tube end than the tube-to-tubesheet weld is.
  • an ultrasonic wave indicated by the arrow L 1 an echo coming from the tube surface is measured.
  • an ultrasonic wave indicated by the arrow L 4 is reflected by a tube surface of a part of the tube-to-tubesheet weld which part is closer to the far side of the tube.
  • an echo of that ultrasonic wave is measured.
  • the probe may be an array probe constituted by a plurality of array elements.
  • the array probe may be disposed so that a direction of arrangement of the array elements coincides with a direction in which the tube extends. With this, it is possible to effectively inspect the tube-to-tubesheet weld whose width extends in the extending direction of the tube.
  • the array probe may be a matrix array probe constituted by array elements arranged in rows and columns.
  • the ultrasonic testing device 7 With use of the data indicated by the result of the measurement carried out by the probe, the ultrasonic testing device 7 generates an ultrasonic testing image that is an image of the echoes of the ultrasonic waves propagated in the tube and the tube-to-tubesheet weld.
  • FIG. 2 illustrates an ultrasonic testing image 111 , which is an example of the ultrasonic testing image generated by the ultrasonic testing device 7 .
  • the information processing device 1 may be configured to generate the ultrasonic testing image 111 . In this case, the ultrasonic testing device 7 transmits, to the information processing device 1 , the data indicating the measurement result obtained by the probe.
  • an intensity of a measured echo is presented as a pixel value of each pixel.
  • An image area of the ultrasonic testing image 111 can be divided into a tube area ar 1 corresponding to the tube, a welded area art corresponding to the tube-to-tubesheet weld, and peripheral echo areas ar 3 and ar 4 where echoes from peripheral parts of the tube-to-tubesheet weld appear.
  • the ultrasonic wave propagated from the probe in a direction indicated by the arrow L 1 is reflected by the tube surface of the part closer to the tube end than the tube-to-tubesheet weld is. This ultrasonic wave is also reflected by the inner surface of the tube. These reflections occur repeatedly. Thus, repetition of echoes a 1 to a 4 appears in the peripheral echo area ar 3 , which extends along the arrow L 1 in the ultrasonic testing image 111 .
  • the ultrasonic wave propagated from the probe in a direction indicated by the arrow L 4 is repeatedly reflected by the outer surface and the inner surface of the tube.
  • the ultrasonic wave propagated from the probe in a direction indicated by the arrow L 3 is not reflected by anything. Thus, no echo appears in an area extending along the arrow L 3 in the ultrasonic testing image 111 . Meanwhile, the ultrasonic wave propagated from the probe in a direction indicated by the arrow L 2 is reflected by the gap, i.e., the defect portion in the tube-to-tubesheet weld. Thus, an echo a 5 appears in an area extending along the arrow L 2 in the ultrasonic testing image 111 .
  • the information processing device 1 analyzes such an ultrasonic testing image 111 to inspect whether or not the tube-to-tubesheet weld has a defect (details thereof will be described later). If the information processing device 1 determines that the tube-to-tubesheet weld has a defect, the information processing device 1 automatically determines the type of the defect, too.
  • FIG. 1 is a block diagram illustrating an example of a configuration of a main part of the information processing device 1 .
  • the information processing device 1 includes a control section 10 which comprehensively controls the sections of the information processing device 1 and a storage section 11 in which various data used by the information processing device 1 is stored.
  • the information processing device 1 further includes an input section 12 which accepts an input manipulation on the information processing device 1 and an output section 13 through which the information processing device 1 outputs data.
  • the control section 10 includes an inspection image generating section 101 , a determining section 102 A, a determining section 102 B, a determining section 102 C, a reliability determining section 103 , a comprehensive determination section 104 , a heat map generating section 105 , a defect type determining section 106 , a thickness calculating section 107 , an integrative detection section 108 , and a defect length calculating section 109 .
  • the storage section 11 has the ultrasonic testing image 111 and inspection result data 112 stored therein.
  • each of the determining sections 102 A, 102 B, and 102 C will be referred to simply as a determining section 102 , in a case where there is no need to distinguish the determining sections 102 A, 102 B, and 102 B from each other.
  • the inspection image generating section 101 cuts an inspection target area from the ultrasonic testing image 111 , so as to generate an inspection image used to determine presence or absence of a defect in the inspection target. A method for generating the inspection image will be described in detail later.
  • the determining section 102 determines a given determination matter in accordance with target data.
  • the target data is an inspection image generated by the inspection image generating section 101
  • the given determination matter is presence or absence of a welding defect in a tube-to-tubesheet weld in a heat exchanger captured in the inspection image.
  • the welding defect may simply be abbreviated as “defect”.
  • the “defect” that is a determination target may be defined in advance in accordance with the purpose and/or the like of the inspection. For example, in a case of a quality inspection of a tube-to-tubesheet weld in a manufactured heat exchanger, it may be determined that a “defect” is present when the inspection image includes an echo caused by a gap inside the tube-to-tubesheet weld or a non-allowable recess on a surface of the tube-to-tubesheet weld. Such a recess is caused by burn-through, for example.
  • the “presence or absence of a defect” can be reworded as presence or absence of a portion (abnormal portion) different from that in a normal product.
  • an abnormal portion detected with use of an ultrasonic waveform or an ultrasonic testing image is generally called “flaw”.
  • the “flaw” is also encompassed in the “defect”.
  • the “defect” further encompasses chipping and cracking.
  • Each of the determining sections 102 A, 102 B, and 102 C determines presence or absence of a defect in accordance with an inspection image generated by the inspection image generating section 101 .
  • the determining methods of the determining sections 102 A, 102 B, and 102 C differ from each other, as will be described later.
  • the determining section 102 A determines presence or absence of a defect with use of a generated image generated by inputting the inspection image into a generative model.
  • the determining section 102 B (numerical analysis/determination section) analyzes pixel values in the inspection image to identify an inspection target portion in the inspection image, and determines presence or absence of a defect in accordance with pixel values in the inspection target portion thus identified.
  • the determining section 102 C determines presence or absence of a defect in accordance with an output value obtained by inputting the inspection image into a decision model. Details of determinations carried out by the determining sections 102 A to 102 C and various models to be used will be described later.
  • the reliability determining section 103 determines a reliability, which is an indicator indicating a degree of certainty of the determination result.
  • the reliabilities are determined in accordance with the inspection image used by the determining sections 102 A to 102 C to derive the determination results (details thereof will be described later).
  • the comprehensive determination section 104 determines presence or absence of a defect with use of (i) the determination results given by the determining sections 102 A to 102 C and (ii) the reliabilities determined by the reliability determining section 103 . With this, it is possible to obtain a determination result in appropriate consideration of the determination results given by the determining sections 102 A to 102 C with a reliability corresponding to the inspection image. Details of the determination method carried out by the comprehensive determination section 104 will be described later.
  • the heat map generating section 105 generates a heat map with use of data obtained in the determining process of the determining section 102 A.
  • the heat map is used by the defect type determining section 106 to determine the type of the defect. The heat map will be described in detail later.
  • the defect type determining section 106 determines the type of the defect included in the inspection image. As discussed above, the determination of the type is carried out with use of the heat map generated by the heat map generating section 105 . A method for determining the type of the defect will be described later.
  • the thickness calculating section 107 calculates a wall thickness of the tube-to-tubesheet weld.
  • the wall thickness calculated by the thickness calculating section 107 can be used as an indicator used to determine whether or not welding has been carried out appropriately. A method for calculating the wall thickness will be described later.
  • the integrative detection section 108 detects, as a single defect, the defects captured in the ultrasonic testing images 111 . The integration of the defect will be described in detail later.
  • the defect length calculating section 109 calculates a length of the defect integrated by the integrative detection section 108 . A method for calculating the length of the defect will be described later.
  • the ultrasonic testing image 111 is an image of an echo of an ultrasonic wave propagated in the inspection target, and is generated by the ultrasonic testing device 7 .
  • the inspection result data 112 refers to data indicating a result of a defect inspection carried out by the information processing device 1 .
  • Recorded in the inspection result data 112 is information indicating a result of determination of presence or absence of a defect, made by the comprehensive determination section 104 , for the ultrasonic testing image 111 stored in the storage section 11 . Further, for an ultrasonic testing image 111 determined as including a defect, the determination result as to the type of the defect made by the defect type determination section 106 is recorded in the inspection result data 112 .
  • information indicating the defect integrated by the integrative detection section 108 is recorded in the inspection result data 112 .
  • FIG. 3 is a view illustrating an outline of an inspection carried out by the information processing device 1 .
  • FIG. 3 shows a process to be carried out after the ultrasonic testing image 111 generated by the ultrasonic testing device 7 is stored in the storage section 11 of the information processing device 1 .
  • the inspection image generating section 101 extracts an inspection target area from the ultrasonic testing image 111 to generate an inspection image 111 A.
  • the extraction of the inspection target area may be carried out with use of an extraction model constructed by machine learning.
  • the extraction model will be described with reference to FIG. 12 .
  • the inspection target area refers to an area sandwiched between two peripheral echo areas ar 3 and ar 4 in each of which an echo coming from the periphery of an inspection target portion of the inspection target appears repeatedly.
  • a given echo caused by the shape and/or the like of the peripheral part is repeatedly observed (echoes a 1 to a 4 and echoes a 6 to a 9 ).
  • the determining sections 102 A, 102 B, and 102 C determine presence or absence of a defect in accordance with the inspection image 111 A. The content of the determination will be described in detail later.
  • the reliability determining section 103 determines reliabilities of the determination results given by the determining sections 102 A, 102 B, and 102 C. Specifically, the reliability of the determination result given by the determining section 102 A is determined in accordance with an output value obtained by inputting the inspection image 111 A into a reliability prediction model for the determining section 102 A. Similarly, the reliability of the determination result given by the determining section 102 B is determined in accordance with an output value obtained by inputting the inspection image 111 A into a reliability prediction model for the determining section 102 B. The reliability of the determination result given by the determining section 102 C is determined in accordance with an output value obtained by inputting the inspection image 111 A into a reliability prediction model for the determining section 102 C.
  • the comprehensive determination section 104 comprehensively determines presence or absence of a defect in accordance with (i) the determination results given by the determining sections 102 A, 102 B, and 102 C and (ii) the reliabilities of the determination results determined by the reliability determining section 103 , and outputs a result of the comprehensive determination. This result is added to the inspection result data 112 .
  • the comprehensive determination section 104 may cause the output section 13 to output the result of the comprehensive determination.
  • the determination result given by the determining section 102 may be represented as a numerical value, and the reliability determined by the reliability determining section 103 may be used as a weight. For example, if the determining sections 102 A, 102 B, and 102 C determine that a defect is present, “1” is output as the determination result. Meanwhile, if the determining sections 102 A, 102 B, and 102 C determine that a defect is absent, “ ⁇ 1” is output as the determination result.
  • the reliability determining section 103 outputs reliabilities within a numerical range from 0 to 1 (a value closer to 1 indicates a higher reliability).
  • the comprehensive determination section 104 may calculate a total value obtained by summing up values obtained by multiplying (i) the values (“1” or “ ⁇ 1”) output by the determining sections 102 A, 102 B, and 102 C by (ii) the reliabilities output by the reliability determining section 103 . Then, the comprehensive determination section 104 may determine presence or absence of a defect in accordance with whether or not the total value thus calculated is higher than a given threshold.
  • the threshold is set at “0”, which is intermediate between “1” indicating that a defect is present and “ ⁇ 1” indicating that a defect is absent.
  • the output values of the determining sections 102 A, 102 B, and 102 C are respectively “1”, “ ⁇ 1”, and “1” and the reliabilities thereof are respectively “0.87”, “0.51”, and “0.95”.
  • the comprehensive determination section 104 carries out calculation as follows: 1 ⁇ 0.87+( ⁇ 1) ⁇ 0.51+1 ⁇ 0.95. The result of the calculation is 1.31, which is higher than “0”, i.e., the threshold. Thus, the result of the comprehensive determination made by the comprehensive determination section 104 indicates that a defect is present.
  • the determining section 102 determines that a defect is present and an echo caused by a welding defect appears at a location above a center position in the image area of the inspection image 111 A, it is considered that the determination result is highly likely to be correct.
  • an area where an echo caused by a welding defect often appears may be preliminarily set. Then, if an echo caused by a welding defect is detected in that area when the determining section 102 determines that a defect is present, the reliability determining section 103 may increase the reliability of the determination result. By correcting, in this manner, the reliability in consideration of the appearance tendency and/or characteristics of the defect, it is possible to make the reliability more appropriate.
  • an area above a center of the image area of the inspection image 111 A may be set as the above-described area. Then, if an echo caused by a welding defect is detected in that area, the reliability determining section 103 may add a given constant to the reliability calculated with use of the reliability prediction model. Meanwhile, if the echo caused by the welding defect is detected at a position outside the above-described area, the reliability determining section 103 may subtract a given constant from the reliability calculated with use of the reliability prediction model.
  • addition of a constant to a reliability is preferably carried out so that the reliability after the addition does not exceed 1.
  • subtraction of a constant from a reliability is preferably carried out such that the reliability after the subtraction is not below 0.
  • the method for correcting the reliability is not limited to the above example.
  • the image area of the inspection image 111 A may be divided into much smaller areas.
  • a value to be added to the reliability may be set higher, as a position of an echo caused by a welding defect is in an area where a welding defect appears with a higher frequency.
  • a value proportional to a distance between a position of a detected echo and a position where a welding defect appears with a highest frequency may be added to the reliability.
  • a value inversely proportional to that distance may be subtracted from the reliability.
  • the reliability may be corrected in consideration of an element(s) other than the position. For example, even in a case where it is determined that a defect is present and an echo caused by a welding defect is detected, it is suspected that the determination result may be incorrect if a pixel value of the echo is low. Thus, as the pixel value of the echo caused by the welding defect is lower, the value of the reliability may be corrected to be lower. Meanwhile, as the pixel value of the echo caused by the welding defect is higher, the value of the reliability may be corrected to be higher.
  • the above-described correction of the reliability can be suitably applied to the determination results given by the determining sections 102 A and 102 B.
  • This is due to the following reason. That is, firstly, in the process in which the determining section 102 A makes determination, a difference image is calculated, and the difference image can be used to calculate the position and/or pixel value of the echo caused by the welding defect. Secondly, in the process in which the determining section 102 B makes determination, the echo caused by the welding defect is detected, and the detection result given thereby can be used.
  • the determining section 102 A determines presence or absence of a defect with use of a generated image obtained by inputting an inspection image into a generative model.
  • the generative model is constructed by machine learning that uses, as training data, an image of an inspection target in which a defect is absent, so that the generative model generates a new image having a similar feature to that of an image input into the generative model.
  • the “feature” is any information obtained from an image. For example, a distribution state, a variance, and the like of pixel values in the image are also included in the “feature”.
  • the generative model is constructed by machine learning that uses, as training data, an image of an inspection target in which a defect is absent.
  • an image of an inspection target in which a defect is absent is input into the generative model as the inspection image, it is highly likely that a new image having a similar feature to that of the inspection image is output as a generated image.
  • the generated image generated from the inspection image in which a defect is captured and (ii) the generated image generated from the inspection image in which no defect is captured differ from each other in that one does not properly restore the target image input into the generative model and the other properly restores the target image input into the generative model.
  • the information processing device 1 that carries out comprehensive determination in consideration of the determination result given by the determining section 102 A that determines presence or absence of a defect with use of the generated image generated by the generative model, it is possible to determine, with high accuracy, presence or absence of a defect having irregular position, size, shape, and/or the like.
  • FIG. 4 is a view illustrating an example of a configuration of the determining section 102 A and an example of a method for determination, carried out by the determining section 102 A, of presence or absence of a defect.
  • the determining section 102 A includes an inspection image obtaining section 1021 , a restored image generating section 1022 , and a defect presence/absence determination section 1023 .
  • the inspection image obtaining section 1021 obtains an inspection image.
  • the information processing device 1 includes the inspection image generating section 101 .
  • the inspection image obtaining section 1021 obtains the inspection image generated by the inspection image generating section 101 .
  • the inspection image may be generated by another device. In this case, the inspection image obtaining section 1021 obtains the inspection image generated by the another device.
  • the restored image generating section 1022 inputs, into the generative model, the inspection image obtained by the inspection image obtaining section 1021 , so as to generate a new image having a similar feature to that of the inspection image thus input.
  • the image generated by the restored image generating section 1022 is called “restored image”.
  • the generative model used to generate the restored image is also called “autoencoder”, and is constructed by machine learning that uses, as training data, an image of an inspection target in which a defect is absent (details thereof will be described later).
  • the defect presence/absence determining section 1023 determines whether or not the inspection target has a defect with use of the restored image generated by the restored image generating section 1022 . Specifically, the defect presence/absence determining section 1023 determines that the inspection target has a defect, if a variance of pixel-by-pixel difference values between the inspection image and the restored image exceeds a given threshold.
  • the inspection image obtaining section 1021 In the method for determining, by the determining section 102 A configured as above, presence or absence of a defect, the inspection image obtaining section 1021 first obtains the inspection image 111 A. Then, the inspection image obtaining section 1021 transmits the obtained inspection image 111 A to the restored image generating section 1022 . As discussed above, the inspection image 111 A is an image generated by the inspection image generating section 101 from the ultrasonic testing image 111 .
  • the restored image generating section 1022 inputs the inspection image 111 A into the generative model, so as to generate a restored image 111 B in accordance with a resulting output value.
  • a method for generating the generative model will be described later.
  • the inspection image obtaining section 1021 removes the peripheral echo areas from the inspection image 111 A to generate a removed image 111 C, and removes the peripheral echo areas from the restored image 111 B to generate a removed image (restored) 111 D.
  • the positions and sizes of the peripheral echo areas captured in the inspection image 111 A are substantially the same, provided that the same inspection target is captured.
  • the inspection image obtaining section 1021 may remove, as a peripheral echo area, a given range in the inspection image 111 A.
  • the inspection image obtaining section 1021 may analyze the inspection image 111 A to detect the peripheral echo areas, and may remove the peripheral echo areas in accordance with a detection result.
  • the defect presence/absence determining section 1023 determines presence or absence of a defect, with respect to a remaining image area obtained by removing the peripheral echo areas from the image area of the restored image 111 B. Consequently, it is possible to determine presence or absence of a defect, without being affected by an echo coming from the periphery. This makes it possible to improve the accuracy in determination of presence or absence of a defect.
  • the defect presence/absence determining section 1023 determines presence or absence of a defect. Specifically, the defect presence/absence determining section 1023 first calculates, in pixels, a difference between the removed image 111 C and the removed image (restored) 111 D. Next, the defect presence/absence determining section 1023 calculates a variance of the difference thus obtained. Then, the defect presence/absence determining section 1023 determines presence or absence of a defect in accordance with whether or not the value of the variance thus calculated exceeds a given threshold.
  • a difference value calculated for a pixel in which an echo caused by a defect appears is higher than difference values calculated for the other pixels.
  • a variance of difference values calculated for a removed image 111 C and a removed image (restored) 111 D generated from an inspection image 111 A where an echo caused by a defect is captured is large.
  • a variance of difference values calculated for a removed image 111 C and a removed image (restored) 111 D generated from an inspection image 111 A where an echo caused by a defect is not captured is relatively small. Meanwhile, in the case where the echo caused by the defect is not captured, a part having somewhat high pixel values due to the effects of noises and/or the like can occur, but a part having extremely high pixel values occurs with low probability.
  • the increase in variance of difference values is a phenomenon characteristic to a case where the inspection target has a defect.
  • the defect presence/absence determining section 1023 configured to determine that a defect is present if a variance of difference values exceeds a given threshold, it is possible to appropriately determine presence or absence of a defect.
  • the defect type determining section 106 determines the type of the defect in accordance with the values of the difference in the pixels calculated by the defect presence/absence determining section 1023 . Since the values of the difference in the pixels indicate the difference between the removed image 111 C and the removed image (restored) 111 D, these values are also called “difference image”.
  • a timing to remove the peripheral echo areas is not limited to the above-described example.
  • a difference image between the inspection image 111 A and the restored image 111 B may be generated, and the peripheral echo areas may be removed from the difference image.
  • the determining section 102 B analyzes pixel values in the inspection image, which is an image of the inspection target, to identify an inspection target portion in the inspection image, and determines presence or absence of a defect in accordance with pixel values in the inspection target portion thus identified.
  • an inspector visually carries out a process of identifying an inspection target portion in an image and checking, in the identified portion, for a defect such as a damage and/or a gap that should not exist from a design standpoint.
  • a visual inspection is requested to be automated, from the viewpoints of reduction of labor, achievement of stable accuracy, and/or the like.
  • the determining section 102 B analyzes the pixel values in the image to identify an inspection target portion in the image, and determines presence or absence of a defect in accordance with pixel values in the inspection target portion thus identified. Thus, it is possible to automate the above-described visual inspection. Further, the information processing device 1 makes determination by comprehensively considering the determination result given by the determining section 102 B and the determination result(s) given by other determining section(s) 102 . Therefore, it is possible to determine, with high accuracy, presence or absence of a defect.
  • the determining section 102 B identifies, as the inspection target portion, an area sandwiched between two peripheral echo areas (peripheral echo areas ar 3 and ar 4 in the example shown in FIG. 2 ) in each of which an echo coming from the periphery of the inspection target portion appears repeatedly. Then, the determining section 102 B determines presence or absence of a defect in accordance with whether or not the identified inspection target portion includes an area (also called “defect area”) constituted by pixel values not less than a threshold.
  • an area also called “defect area”
  • the determining section 102 B may first binarize the inspection image 111 A with use of a given threshold to generate a binarized image. Then, the determining section 102 B detects the peripheral echo areas from the binarized image.
  • the inspection image 111 A shown in FIG. 3 includes echoes a 1 , a 2 , a 6 , and a 7 .
  • the determining section 102 B can detect these echoes in the binarized image. Then, the determining section 102 B can detect edges of these echoes thus detected, and can identify, as the inspection target portion, an area surrounded by these edges.
  • the determining section 102 B identifies a right edge of the echo a 1 or a 2 as a left edge of the inspection target portion, and identifies a left edge of the echo a 6 or a 7 as a right edge of the inspection target portion. These edges constitute boundaries between (i) the peripheral echo areas ar 3 and ar 4 and (ii) the inspection target portion. Similarly, the determining section 102 B identifies an upper edge of the echo a 1 or a 6 as an upper edge of the inspection target portion, and identifies a lower edge of the echo a 2 or a 7 as a lower edge of the inspection target portion.
  • the determining section 102 B may set the upper edge of the inspection target portion at a location above the upper edge of the echo a 1 or a 6 .
  • the determining section 102 B can analyze the inspection target portion identified in the binarized image to determine whether or not the echo caused by the defect is captured therein. For example, in a case where the inspection target portion includes a continuous area constituted by a given number or more of pixels, the determining section 102 B may determine that the echo caused by the defect is captured at a location where the continuous area exists.
  • the determining section 102 B may determine presence or absence of a defect in accordance with the value of the variance.
  • the determining section 102 C determines presence or absence of a defect in accordance with an output value obtained by inputting the inspection image into a decision model.
  • the decision model is constructed by, e.g., carrying out machine learning with use of (i) training data generated by using an ultrasonic testing image 111 of an inspection target in which a defect is present and (ii) training data generated by using an ultrasonic testing image 111 of an inspection target in which a defect is absent.
  • the decision model can be constructed by any learning model suitable for image classification.
  • the decision model may be constructed by, e.g., convolutional neural network having excellent image classification accuracy.
  • FIG. 5 is a view illustrating an example in which a heat map is generated from an ultrasonic testing image and threshold processing is carried out on the heat map thus generated.
  • the upper part of FIG. 5 illustrates an example of an ultrasonic testing image 111 - a of a portion of a tube-to-tubesheet weld in which a defect is present
  • the lower part of FIG. 5 illustrates an example of an ultrasonic testing image 111 - b of a portion of the tube-to-tubesheet weld in which a defect is absent.
  • an inspection image 111 A is generated from the ultrasonic testing image 111 , and a restored image 111 B is generated from the inspection image 111 A. Then, a removed image 111 C is generated from the inspection image 111 A, and a removed image (restored) 111 D is generated from the restored image 111 B.
  • a removed image 111 C-a and a removed image (restored) 111 D-a are generated from the ultrasonic testing image 111 - a .
  • a difference image is generated from the removed image 111 C-a and the removed image (restored) 111 D-a.
  • the heat map generating section 105 generates a heat map in which pixels in the difference image are expressed by colors or gradations corresponding to the pixel values.
  • FIG. 5 shows a heat map 111 E-a in which pixel values, from the lower limit to the higher limit, are expressed by gradations of colors from black to white.
  • an area corresponding to the defect i.e., an area in which pixels having high pixel values are collected
  • the area corresponding to the defect can be easily visually recognized.
  • the heat map 111 E-a also has an area in which pixel values have become high due to noises and/or the like.
  • the heat map generating section 105 carry out the threshold processing on the heat map thus generated, so as to correct the pixel values in the area in which the pixel values have become high due to noises and/or the like.
  • the heat map generating section 105 may set, at zero (black), a pixel value(s) in the heat map 111 E-a which pixel value(s) is/are not more than a given threshold. Consequently, a heat map 111 F-a from which a noise component(s) has/have been removed is generated. With the heat map 111 F-a, the area corresponding to the defect can be more clearly recognized.
  • an ultrasonic testing image 111 - b of a portion in which a defect is absent A removed image 111 C-b and a removed image (restored) 111 D-b are generated from the ultrasonic testing image 111 - b , and a difference image is generated from the removed image 111 C-b and the removed image (restored) 111 D-b.
  • the heat map generating section 105 generates a heat map 111 E-b of the difference image, and carries out the threshold processing on the heat map 111 E-b to generate a heat map 111 F-b. It is understood that comparison of the heat map 111 F-a with the heat map 111 F-b makes it possible to clearly determine presence or absence of a defect. It is also understood that the position of the defect can be clearly identified in the heat map 111 F-a.
  • the incomplete penetration in the first layer refers to creation of a gap due to incomplete welding occurred in the vicinity of the tubesheet.
  • the incomplete fusion between the welding passes refers to creation of a gap due to poor welding occurred while carrying out welding plural times.
  • the undercut refers to such a defect that an end of a weld bead is hollowed out in the form of a notch.
  • the blowhole refers to a spherical void formed in the welding metal.
  • the positions where these defects occur differ from each other.
  • the type of the defect in accordance with the position in the ultrasonic testing image 111 at which position an echo caused by the defect appears, it is possible to determine the type of the defect.
  • the type of the defect in accordance with the position of the defect area in the heat map (preferably, the one after the threshold processing) generated from the ultrasonic testing image 111 .
  • the defect area is an area where an echo caused by the defect appears, and therefore pixel values therein are higher than those in the other areas.
  • FIG. 6 is a view illustrating a relation between the position of the defect, the ultrasonic testing image, and the heat map.
  • the view at the left end in the first row in FIG. 6 shows a cross section of a tube-to-tubesheet weld in which incomplete penetration in the first layer occurs.
  • the left side in FIG. 6 is the tube end side, whereas the right side in FIG. 6 is the far side of the tube. That is, the tube extends in a left-right direction in FIG. 6 .
  • the tubesheet is located on a lower side relative to the tube wall.
  • a scale is put on an inner wall surface (inner surface) of the tube so as to indicate a width of the tube-to-tubesheet weld.
  • an area indicated by the broken line is a welded area of the tubesheet obtained by welding, an inverted triangle area on the left side of the welded area is an area made of a welding metal, and an area obtained by combining these areas is the tube-to-tubesheet weld.
  • a gap is created in the circled portion of the tube-to-tubesheet weld. This gap is positioned at a location that is in the vicinity of the surface of the tube and that is closer to the end of the tube-to-tubesheet weld which end is closer to the far side of the tube.
  • an echo caused by the gap appears in an ultrasonic testing image 111 - c of the portion including the gap.
  • an area corresponding to the gap appears also in a heat map 111 F-c generated from the ultrasonic testing image 111 - c.
  • the view at the left end in the second row in FIG. 6 shows a cross section of a tube-to-tubesheet weld in which incomplete fusion between welding passes occurs.
  • a gap is created in the circled portion. This gap is positioned at a location that is in the vicinity of the surface of the tube and that is near a center part in a thickness direction of the tube-to-tubesheet weld.
  • an echo caused by the gap appears in an ultrasonic testing image 111 - d of the portion including the gap.
  • an area corresponding to the gap appears also in a heat map 111 F-d generated from the ultrasonic testing image 111 - d . This area is located more leftward than that in the heat map 111 F-c on the first row.
  • the view at the left end in the third row in FIG. 6 shows a tube-to-tubesheet weld in which undercut occurs, viewed from the tube end side.
  • a gap is created. This gap is positioned at a location that is in the vicinity of the surface of the tube and that is near an end of the tube-to-tubesheet weld which end is closer to the tube end side.
  • an echo caused by the gap appears in an ultrasonic testing image 111 - e of the portion including the gap.
  • an area corresponding to the gap appears also in a heat map 111 F-e generated from the ultrasonic testing image 111 - e . This area is located more leftward than that in the heat map 111 F-d on the second row.
  • the view at the left end in the fourth row in FIG. 6 shows a cross section of a tube-to-tubesheet weld in which a blowhole is created.
  • a gap is created in the circled portion. This gap is positioned at a location that is closer to the inside of the tube-to-tubesheet weld rather than to the surface of the tube, and the position of this gap in a left-right direction is in the vicinity of a center of the tube-to-tubesheet weld in its width direction.
  • an echo caused by the gap appears in an ultrasonic testing image 111 - f of the portion including the gap.
  • an area corresponding to the gap appears also in a heat map 111 F-f generated from the ultrasonic testing image 111 - f .
  • the position of this area in the left-right direction is close to that in the heat map 111 F-d in the second row.
  • the position of this area in an up-down direction is more downward than that in the heat map 111 F-d in the second row.
  • a type decision model used to determine the type of the defect can be constructed by machine learning that uses, as training data, a heat map of a difference image generated from an inspection image of an inspection target having a defect of a known type. Then, the defect type determining section 106 can determine the type of the defect in accordance with an output value obtained by inputting, into such a decision model, a heat map generated by the heat map generating section 105 .
  • a heat map representing, by colors or gradations, pixel values of pixels constituting a difference image reflects a difference in the type of the defect captured in an inspection image from which the difference image is obtained.
  • a large number of heat maps like the heat map 111 F-c shown in FIG. 6 which is generated from the ultrasonic testing image 111 of the portion in which incomplete penetration in the first layer occurs, may be prepared and used as training data. With this, it is possible to construct a type decision model that outputs a probability that the type of a defect is incomplete penetration in the first layer. Similarly, carrying out machine learning by using, as training data, a heat map generated from an ultrasonic testing image 111 of a portion in which another type of defect occurs makes it possible to construct a type decision model that outputs probabilities of various types of defects.
  • the defect type determining section 106 can determine the type of the defect in accordance with the output value obtained by inputting a heat map into such a type decision model. For example, the defect type determining section 106 may determine that a defect of the type corresponding to, among the probability values corresponding to various types of defects output from the type decision model, a highest probability value occurs.
  • the defect type determining section 106 detects a defect area from a difference image, and determines the type of the defect related to the defect area, in accordance with the position in an image area of the difference image at which position the defect area is detected.
  • FIG. 7 is a view illustrating the method for detecting the defect area.
  • FIG. 7 illustrates an example in which a defect area is detected with use of a heat map.
  • the generation of the heat map is not essential, as discussed below.
  • FIG. 7 shows (i) a heat map 111 E generated from an ultrasonic testing image 111 of an inspection target in which a defect is present and (ii) a heat map 111 F obtained by carrying out threshold processing on the heat map 111 E.
  • FIG. 7 also shows an enlarged view of the upper left part of the heat map 111 F. In this enlarged view, the pixels in the heat map 111 F indicate their pixel values.
  • the defect type determining section 106 In order to detect the defect area, the defect type determining section 106 first detects, in the heat map 111 F, a pixel having a highest pixel value. In the example shown in FIG. 7 , the highest pixel value is 104. Thus, the pixel having this pixel value is detected. Next, the defect type determining section 106 detects a pixel(s) being adjacent to the detected pixel and having a pixel value(s) not less than a given threshold.
  • the defect type determining section 106 carries out such a process repeatedly until no adjacent pixel having a pixel value not less than the threshold is detected. Consequently, the defect type determining section 106 can detect, as a defect area, a continuous area constituted by the pixels each having a pixel value not less than the given threshold.
  • the defect type determining section 106 may detect, as the defect area, a quadrangular area ar 5 including the defect area detected in the above-described manner.
  • the above-discussed process can be carried out, if there is data indicating pixel-by-pixel difference values, i.e., a difference image between the inspection image 111 A and the restored image 111 B.
  • a difference image between the inspection image 111 A and the restored image 111 B.
  • the defect type determining section 106 detects, as a defect area, an area constituted by a plurality of pixels which are included in a difference image and which have pixel values not less than a threshold.
  • the pixel values of the pixels in the defect area are higher than the pixel values in the other areas.
  • the defect type determining section 106 determines the type of a defect related to a defect area, in accordance with the position in the image area of the difference image at which position the defect area is detected. With this, it is possible to automatically determine the type of the defect.
  • the defect type determining section 106 can determine the type of the defect in accordance with which of the areas the defect area detected in the above-described manner is contained.
  • FIG. 8 is a view illustrating an example of areas set for respective types of defect.
  • an area AR 1 corresponding to undercut is set at the upper left corner
  • an area AR 2 corresponding to incomplete fusion between welding passes is set at the upper center
  • an area AR 3 corresponding to incomplete penetration in the first layer is set at the upper right corner in the heat map 111 F.
  • an area AR 4 corresponding to a blowhole is set at a location slightly above the center.
  • These areas may be set in advance in accordance with, e.g., analysis on a difference image and/or a heat map that is/are based on inspection images of portions in which various defects are present.
  • the defect area indicated by the hollow arrow is detected in the area AR 3 .
  • the defect type determining section 106 determines that this defect is caused by incomplete penetration in the first layer.
  • a part of the area AR 4 which corresponds to a blowhole, overlaps parts of the areas AR 1 to AR 3 .
  • an area used to determine the type of defect may be set so as to partially overlap another area.
  • the defect type determining section 106 may determine, as a determination result regarding the type of the defect, all the types corresponding to the plurality of areas. For example, when a defect area is detected in an area where the areas AR 1 and AR 4 overlap each other, the defect type determining section 106 may output, as a determination result, both undercut and a blowhole.
  • the defect type determining section 106 may narrow down the determination results of the type of the defect in accordance with whether or not a condition(s) specific to each type of defect is/are satisfied. For example, in a case of a defect that is characterized by shape, a condition related to the shape may be set. Meanwhile, in a case of a defect that is characterized by size, a condition related to the size may be set.
  • a blowhole is such a defect that causes a spherical cavity, a diameter of which is generally not more than 2 mm.
  • a single ultrasonic testing image 111 covers a range of approximately 1 mm of the inspection target in width
  • a single blowhole will fit within two or three ultrasonic testing images 111 or so. Therefore, if a defect is detected consecutively in ultrasonic testing images 111 respectively corresponding to adjacent parts of the inspection target and the number of ultrasonic testing images 111 is not more than three, the defect may possibly be a blowhole. Meanwhile, if the number of ultrasonic testing images 111 in which a defect is detected consecutively is not less than four, it is highly likely that the defect is not a blowhole.
  • the defect type determining section 106 may determine that the type of the defect is a blowhole, if the condition that the number of ultrasonic testing images 111 in which the defect is consecutively detected is not more than a threshold (e.g., 3) is satisfied.
  • the defect type determining section 106 may determine that the type of the defect is a blowhole. Meanwhile, if the number of ultrasonic testing images 111 in which the defect is consecutively detected exceeds the threshold, the defect type determining section 106 may determine that the type of the defect is incomplete fusion between welding passes.
  • a blowhole has a spherical shape.
  • the peak values of the echoes caused by the blowhole observed in the ultrasonic testing images 111 often differ from each other. Such a difference in peak value appears as a difference in pixel value in the ultrasonic testing images 111 . For example, suppose that a single blowhole is detected across three ultrasonic testing images 111 .
  • a peak value of an echo caused by the blowhole observed in the intermediate one of the three ultrasonic testing images 111 is 50%
  • a peak value of an echo caused by the blowhole in each of the ultrasonic testing images 111 before and after the intermediate one is 30%, which is lower than that of the intermediate one.
  • the defect type determining section 106 may determine that the type of the defect is a blowhole, if the condition that there is a difference between the pixel values in the defect areas in the ultrasonic testing images 111 in which the defect is consecutively detected is satisfied. For example, the defect type determining section 106 may calculate average values of the pixel values in the pixels included in the defect areas in the ultrasonic testing images 111 . Then, if a difference between the average values is not less than a threshold, the defect type determining section 106 may determine that there is a difference.
  • the defect type determining section 106 may carry out either or both of (i) determination involving use of the type decision model and (ii) determination in accordance with which of the areas the defect area is contained. Carrying out both of the determinations makes it possible to enhance the accuracy in the determination result regarding the type.
  • the tube-to-tubesheet weld surrounds the tube by 360 degrees.
  • ultrasonic testing images 111 of various parts of the tube-to-tubesheet weld are generated by circumferentially moving, in the tube, the probe by a given degree. Then, in accordance with the ultrasonic testing images 111 , a defect is detected.
  • a single continuous defect is captured in a plurality of ultrasonic testing images and accordingly the defect is determined as two or more defects, although the entity of the defect is a single defect.
  • the integrative detection section 108 integrates the defects captured in the plurality of ultrasonic testing images 111 so as to detect the defects as a single defect.
  • the comprehensive determination section 104 determines that a defect is present in a plurality of ultrasonic testing images 111 corresponding to parts of the tube-to-tubesheet weld which parts are adjacent to each other, the integrative detection section 108 detects, as a single defect, the defects captured in the plurality of ultrasonic testing images 111 . This makes it possible to carry out detection appropriate to the entity of the defect.
  • FIG. 9 is a view illustrating a method for integrating defects captured in a plurality of ultrasonic testing images 111 to detect the defects as a single defect.
  • the upper left part of FIG. 9 illustrates a transverse cross section of a tube and a tube-to-tubesheet weld.
  • the lower left part of FIG. 9 illustrates a longitudinal cross section of the tube, the tube-to-tubesheet weld, and a tubesheet.
  • a welding defect occurs in a wide range along an outer wall surface of the tube.
  • measurement results obtained in a range in which a welding defect occurs reflect echoes from the welding defect. Consequently, as shown in the right part of FIG. 9 , echoes caused by the welding defect appear in ultrasonic testing images 111 g to 111 i generated in accordance with the measurement results.
  • the comprehensive determination section 104 determines that a defect is present.
  • the ultrasonic testing images 111 g to 111 i respectively correspond to parts of the tube-to-tubesheet weld which parts are adjacent to one another.
  • the integrative detection section 108 detects, as a single defect, the defects captured in the ultrasonic testing images 111 g to 111 i that are determined as including a defect by the comprehensive determination section 104 .
  • the integrative detection section 108 may integrate the defects, provided that the positions of the defects detected in the ultrasonic testing images 111 g to 111 i are the same or close to each other. As discussed above, the position of the defect varies depending on the type. Thus, the integrative detection section 108 may integrate these defects, provided that the defects of the same type are detected in the ultrasonic testing images 111 g to 111 i . With these configurations, it is possible to enhance the accuracy in defect integration.
  • the defect length calculating section 109 calculates a length of the defect integrated by the above-described processing.
  • the defect length calculating section 109 may calculate the length of the defect by multiplying (i) a length of a defect per ultrasonic testing image 111 by (ii) the number of defects integrated by the integrative detection section 108 .
  • a defect captured in a single ultrasonic testing image 111 has a length of approximately (outer diameter of tube) ⁇ 1/360.
  • the defect length calculating section 109 may calculate the length of the defect as follows: (outer diameter of tube) ⁇ 3 ⁇ 1/360. Note that ⁇ denotes pi.
  • FIG. 10 is a view illustrating the method for calculating the thickness of the tube-to-tubesheet weld.
  • the lower part of FIG. 10 illustrates a longitudinal cross section of the tube-to-tubesheet weld
  • the upper part of FIG. 10 illustrates an ultrasonic testing image 111 of the tube-to-tubesheet weld.
  • a thickness of the tube-to-tubesheet weld shown in the lower part of FIG. 10 , including a part penetrated into the tubesheet, is X.
  • the area of the ultrasonic testing image 111 in which area the tube-to-tubesheet weld is captured is an area sandwiched between two peripheral echo areas ar 3 and ar 4 in each of which an echo coming from the periphery of the peripheral echo area appears repeatedly.
  • the thickness X of the tube-to-tubesheet weld can be calculated on the basis of a distance Xi between the peripheral echo areas ar 3 and ar 4 .
  • the distance Xi may be calculated by the thickness calculating section 107 analyzing the ultrasonic testing image 111 .
  • the analysis of the ultrasonic testing image 111 is carried out by the determining section 102 B, it is preferable to use the result of the analysis.
  • the determining section 102 B detects the peripheral echo areas ar 3 and ar 4 in the inspection image 111 A generated from the ultrasonic testing image 111 , and carries out, with respect to an area sandwiched between these areas, determination of presence or absence of a defect.
  • the thickness calculating section 107 can calculate the distance Xi from the right edge of the peripheral echo area ar 3 to the left edge of the peripheral echo area ar 4 detected by the determining section 102 B. Further, if a reduced scale of the inspection image 111 A is obtained in advance, the thickness calculating section 107 can use the reduced scale to calculate the thickness X of the tube-to-tubesheet weld.
  • the determining section 102 B detects the peripheral echo areas ar 3 and ar 4 in the process of determining presence or absence of a defect.
  • the thickness calculating section 107 can use the result of detection of the peripheral echo areas ar 3 and ar 4 carried out by the determining section 102 B to calculate the thickness of the inspection target portion.
  • the determination result of presence or absence of a defect in the inspection target made by the information processing device 1 is output via the output section 13 .
  • an example of output of an inspection result will be explained with reference to FIG. 11 .
  • FIG. 11 shows an example of output of an inspection result.
  • the upper left part of FIG. 11 shows a defect map 300 .
  • the defect map 300 includes a doughnut-shaped area 301 showing a tube-to-tubesheet weld viewed from the tube end side, with line segments 302 indicating detected defects drawn.
  • the defect map 300 can facilitate recognition of the distribution of the defects in the tube-to-tubesheet weld.
  • the upper right part of FIG. 11 shows a tubesheet map 400 .
  • the tubesheet map 400 schematically shows a state of a heat exchanger including a tubesheet to which many tubes are welded as shown in FIG. 2 , viewed from the tube end side.
  • the tubesheet map 400 shows inspection results by drawing, at the location of each tube, a graphic indicating a result of a defect inspection carried out at a tube-to-tubesheet weld of that tube.
  • a white circle is drawn at the position of a tube where no defect was detected as a result of the inspection
  • a black circle is drawn at the position of a tube where a flaw (defect) was detected as a result of the inspection.
  • This can facilitate recognition of the distribution of the tube-to-tubesheet welds where defects have occurred.
  • a triangle is drawn at the position of a tube which has not been inspected yet, and a square is drawn at the position of a tube which is not a target to be inspected. In this manner, various information regarding inspections may also be included in the tubesheet map 400 .
  • the lower part of FIG. 11 shows an ultrasonic testing image set 500 .
  • the ultrasonic testing image set 500 includes three ultrasonic testing images ( 501 to 503 ).
  • the ultrasonic testing image 501 is obtained by sector scanning on the tube end side
  • the ultrasonic testing image 502 is obtained by linear scanning
  • the ultrasonic testing image 503 is obtained by sector scanning on the far side of the tube.
  • the linear scanning is scanning carried out in a flaw detecting direction perpendicular to a center axis of the tube.
  • the above-described ultrasonic testing image 111 is also obtained by linear scanning.
  • the sector scanning on the tube end side is scanning with which an ultrasonic wave is propagated in a flaw detecting direction that is inclined toward the far side of the tube from the direction perpendicular to the center axis of the tube.
  • the sector scanning on the far side of the tube is scanning with which an ultrasonic wave is propagated in a flaw detecting direction that is inclined toward the tube end side from the direction perpendicular to the center axis of the tube.
  • each of these ultrasonic testing images a reflected echo corresponding to the detected defect is marked. Indicating the ultrasonic testing images with the markings as inspection results in this manner can facilitate recognition of the position and/or the like of the defects.
  • the information processing device 1 determines presence or absence of a defect in a plurality of ultrasonic testing images 111 obtained by scanning in different flaw detecting directions. Then, if the information processing device 1 determines that a defect is present in any of the flaw detecting directions, the information processing device 1 may determine, as a final determination result, that a defect is present even when no defect is found in the other flaw detecting directions. This can reduce the probability that a defect is missed. Alternatively, the information processing device 1 may determine presence or absence of a defect, with respect to a composite image that is a composite of (i) an ultrasonic testing image obtained by linear scanning and (ii) an ultrasonic testing image obtained by sector scanning.
  • the information processing device 1 may output, as an inspection result, all of or only apart of the defect map 300 , the tubesheet map 400 , and the ultrasonic testing image set 500 .
  • the information processing device 1 may also output, as an inspection result, information indicating the determination result regarding the type of the defect, for example. Needless to say, these are merely examples.
  • the information processing device 1 may output a determination result in any form with which a person can recognize the content thereof.
  • FIG. 12 is a view illustrating an example of the process of constructing various models used in an inspection and determining a threshold. These processes may be carried out by the information processing device 1 or by another computer.
  • ultrasonic testing images 111 having been subjected to smoothing processing are obtained.
  • the ultrasonic testing images 111 include ultrasonic testing images 111 obtained from inspection targets in which a defect is present and ultrasonic testing images 111 obtained from inspection targets in which a defect is absent.
  • the ultrasonic testing images 111 obtained from the inspection targets in which a defect is present are classified in accordance with their types of defects.
  • the smoothing processing refers to a process of smoothing a change in pixel value between pixels adjacent to each other.
  • the smoothing processing may be carried out either by the information processing device 1 or the ultrasonic testing device 7 .
  • the smoothing processing is not essential. However, it is preferable to carry out the smoothing processing, since carrying out the smoothing processing makes it easier to distinguish an echo caused by a defect from a noise component.
  • an extraction model is constructed.
  • the extraction model is constructed by machine learning that uses training data in which an ultrasonic testing image 111 is associated with extraction area information as correct data.
  • the extraction area information is information indicating an area to be extracted from the ultrasonic testing image 111 , that is, information indicating an inspection target area.
  • the extraction area information may be the one generated in the following manner. That is, for example, the ultrasonic testing image 111 is displayed by a display device, an operator inputs an area to be extracted, and the extraction area information is generated in accordance with the content of the input.
  • the extraction model can be constructed by any learning model suitable for extraction of an area from an image.
  • the extraction model may be constructed by You Only Look Once (YOLO) or the like that involves excellent extraction accuracy and excellent processing speed.
  • the area to be extracted can be any area that includes a tube-to-tubesheet weld, which is an inspection target portion. It is preferable that the area to be extracted also include at least apart of an area where an echo coming from its periphery appears. This is preferable because: if the inspection target portion has no defect, no feature point that can be machine-learned may be observed in that portion of the ultrasonic testing image 111 ; in such a case, it is difficult to construct an extraction model. For example, in the ultrasonic testing image 111 shown in FIG. 2 , an area(s) including parts of the echoes a 1 , a 2 , a 6 , and a 7 may be set as the area to be extracted. With this, it is possible to construct the extraction model that can extract an area which includes the tube-to-tubesheet weld and the echoes coming from the periphery.
  • the extraction model constructed in S 2 is used to generate images for learning (i.e., learning images) from the ultrasonic testing images 111 obtained in S 1 .
  • learning images i.e., learning images
  • this extraction model is used to extract the inspection target area.
  • an echo from the periphery of the inspection target portion has a feature that can be machine-learned. Therefore, with this configuration, it is possible to extract the inspection target portion automatically with high accuracy.
  • the learning images are generated with use of the same extraction model as that used for the inspection image 111 A, the learning images have similar appearances to that of the inspection image 111 A (for the appearance of the inspection image 111 A, see FIG. 3 ).
  • the learning images generated in S 3 are used for determination of a threshold and construction of a model related to the determining sections 102 .
  • a generative model is constructed.
  • the generative model is constructed by machine learning that uses, as training data, the learning images generated from the ultrasonic testing images 111 of the inspection targets in which a defect is absent.
  • the generative model may be an autoencoder.
  • the generative model may be a model obtained by improving or modifying the autoencoder.
  • the generative model may be a variational autoencoder or the like.
  • the extraction model is constructed by machine learning that uses, as correct data, the area including the area where the echo coming from the periphery appears
  • the training data to be used to construct the generative model also includes the area where the echo coming from the periphery appears.
  • the inspection target area does not include any echo, and thus does not have enough feature points to be machine-learned.
  • the training data including the area where the echo coming from the periphery appears it is possible to construct an appropriate generative model.
  • test images are input into the generative model constructed in S 4 to generate a restored image.
  • the test images are, among the learning images generated in S 3 , the ones that are not used to construct the generative model.
  • the test images include test images generated from the ultrasonic testing images 111 of the inspection targets in which a defect is absent and test images generated from the ultrasonic testing images 111 of the inspection targets in which a defect is present.
  • the test images generated from the ultrasonic testing images 111 of the inspection targets in which a defect is present are classified in accordance with their types.
  • a difference between the restored image and the test images is calculated in pixels, and a variance of the difference is calculated.
  • a threshold is defined so that (i) variance values calculated for a plurality of test images generated from the ultrasonic testing images 111 of the inspection targets in which a defect is absent and (ii) variance values calculated for a plurality of test images generated from the ultrasonic testing images 111 of the inspection targets in which a defect is present can be distinguished from each other.
  • a reliability prediction model for the determining section 102 A which carries out determination with use of the generative model constructed in S 4 and the threshold determined in S 5 , is constructed by machine learning.
  • the machine learning uses training data in which a test image is associated with, as correct data, information indicating whether or not a result of determination made by the determining section 102 A in accordance with that test image is correct.
  • the test image may be the one generated from an ultrasonic testing image 111 for which presence or absence of a defect is known.
  • the reliability determining section 103 may use an output value from the reliability prediction model as the reliability of the determination result given by the determination result 102 A.
  • a type decision model is constricted by using, as training data, heat maps generated respectively from test images of various types of defects. As discussed with reference to FIG. 6 , the heat maps show features corresponding to the types of the defects. Thus, by carrying out machine learning with use of the heat maps as training data, it is possible to construct a type decision model.
  • the type decision model can be constructed by any learning model suitable for image classification.
  • the type decision model may be constructed by, e.g., convolutional neural network that has excellent image classification accuracy.
  • areas for type determination are set. Specifically, first, from the heat maps corresponding to the various types of defects and being generated in S 7 , defect areas where echoes caused by the defects are captured are detected. Then, in an image area of each heat map, an area where a defect area of a certain type is detected is identified as an area where the defect of the certain type appears. This process is carried out for each of the types that a user wants to carry out determination. With this, as in the example shown in FIG. 8 , it is possible to set, for each type of defect, an area where that type of defect appears.
  • either of S 7 and S 8 may be omitted. If the process in S 7 is omitted, the defect type determining section 106 determines the type of the defect in accordance with the areas set in S 8 . Meanwhile, if the process in S 8 is omitted, the defect type determining section 106 determines the type of the defect with use of the type decision model constructed in S 7 .
  • a threshold to be used by the determining section 102 B for numerical analysis is determined with use of the learning images generated in S 3 . For example, in a case where the determining section 102 B carries out banalization processing, a threshold used for the banalization processing is determined.
  • a reliability prediction model for the determining section 102 B which carries out determination with use of the threshold determined in S 9 , is constructed by machine learning.
  • the machine learning uses training data in which a test image is associated with, as correct data, information indicating whether or not a result of determination made by the determining section 102 B in accordance with that test image is correct.
  • the test image may be the one generated from an ultrasonic testing image 111 for which presence or absence of a defect is known.
  • a decision model used by the determining section 102 C to determine presence or absence of a defect is constructed by machine learning.
  • the machine learning uses training data in which a learning image generated in S 3 is associated with, as correct data, information indicating presence or absence of a defect. With this, it is possible to construct a decision model that outputs, in response to inputting the inspection image 111 A into the decision model, a value indicating the probability that a defect is present or a value indicating the probability that a defect is absent.
  • a reliability prediction model for the determining section 102 C which carries out determination with use of the decision model constructed in S 11 , is constructed by machine learning.
  • the machine learning uses training data in which a test image is associated with, as correct data, information indicating whether or not a result of determination made by the determining section 102 C in accordance with that test image is correct.
  • the test image may be the one generated from an ultrasonic testing image 111 for which presence or absence of a defect is known.
  • the reliability prediction model for the determining section 102 A can be constructed by machine learning that uses training data in which the test image for which the determining section 102 A has carried out determination for presence or absence of a defect is associated with, as correct data, information indicating whether or not the result of the determination is correct. This is also true of the reliability prediction model for the determining section 102 B and the reliability prediction model for the determining section 102 C.
  • Each of the reliability prediction models described above has learned correspondence between (i) the test image for which the determining section 102 has carried out determination and (ii) the information indicating whether or not the result of the determination is correct.
  • an output value obtained in response to inputting the inspection image 111 A into the reliability prediction model indicates the degree of certainty of a determination result obtained when the determining section 102 carries out determination with use of the inspection image 111 A.
  • the reliability determining section 103 can determine the reliabilities of the determination results of the determining sections 102 in accordance with output values obtained by inputting the inspection image 111 A into the reliability prediction models for the determining sections 102 , and accordingly can set reliabilities appropriate for the previous determination history.
  • FIG. 13 is a view illustrating an example of an inspection method involving use of the information processing device 1 .
  • the storage section 11 stores therein ultrasonic testing images 111 each of which is an image of echoes coming from a tube-to-tubesheet weld and its periphery and measured by circumferentially moving the probe.
  • the inspection image generating section 101 generates the inspection image 111 A.
  • the inspection image generating section 102 obtains one of the ultrasonic testing images 111 stored in the storage section 11 and inputs the one of the ultrasonic testing images 111 into the extraction model, and extracts, from the one of the ultrasonic testing images 111 , an area indicated by a resulting output value so as to generate an inspection image 111 A.
  • the determining sections 102 determine presence or absence of a defect with use of the inspection image 111 A generated in S 21 .
  • the inspection image obtaining section 1021 obtains the inspection image 111 A generated in S 21
  • the restored image generating section 1022 generates a restored image 111 B from the inspection image 111 A with use of the generative model constructed in S 4 in S 12 .
  • the defect presence/absence determining section 1023 calculates pixel-by-pixel differences between the inspection image 111 A and the restored image 111 B, and calculates a variance of the differences.
  • the defect presence/absence determining section 1023 determines presence or absence of a defect in accordance with whether or not the value of the variance is higher than the threshold determined in S 5 in FIG. 12 . Note that, in a case where the removed image 111 C and the removed image (restored) 111 D have been generated, the defect presence/absence determining section 1023 calculates differences between these images.
  • the determining section 102 B binarizes the inspection image 111 A generated in S 21 with use of the threshold determined in S 9 in FIG. 12 , so as to generate a binarized image. Then, the determining section 102 B detects peripheral echo areas ar 3 and ar 4 in the binarized image thus generated, and determines presence or absence of a defect in accordance with whether or not an area sandwiched between these areas includes a defect area.
  • the determining section 102 C inputs the inspection image 111 A generated in S 21 into the decision model constructed in S 11 in FIG. 12 , and determines presence or absence of a defect in accordance with a resulting output value. For example, in a case of using the decision model that outputs the probability that a defect is present, the determining section 102 C may determine that a defect is present if an output value from the decision model exceeds a given threshold. Such a threshold is also determined after the process in S 11 in FIG. 12 .
  • the reliability determining section 103 determines the reliability of the determination result given by the determining section 102 , with use of the inspection image 111 A generated in S 21 . Specifically, the reliability determining section 103 determines the reliability of the determination result given by the determining section 102 A, in accordance with an output value obtained by inputting the inspection image 111 A into the reliability prediction model constructed in S 6 in FIG. 12 .
  • the reliability determining section 103 may employ this value as the reliability as it is. Further, the reliability determining section 103 determines the reliability of the determination results of the determining sections 102 B and 102 C in a similar manner. In this manner, for each of the determining sections 102 A to 102 C, the reliability determining section 103 determines the reliability of the result of determination of presence or absence of a defect.
  • the comprehensive determination section 104 determines presence or absence of a defect with use of the determination results obtained in S 22 and the reliabilities determined in S 23 . Specifically, the comprehensive determination section 104 determines presence or absence of a defect with use of numerical values obtained by summing up the values obtained by weighing, in accordance with their reliabilities, the numerical values indicating the determination results of the determining section 102 A to 102 C.
  • each of the determination results of the determining sections 102 A to 102 C can be expressed by a numerical value “ ⁇ 1” (a defect is absent) or “1” (a defect is present).
  • the determination results may be multiplied by the values of the reliabilities as they are.
  • the comprehensive determination section 104 carries out calculation in accordance with the following expression: 1 ⁇ 0.87+( ⁇ 1) ⁇ 0.51+1 ⁇ 0.95. Consequently, a numerical value of 1.31 is obtained.
  • the comprehensive determination section 104 compares this numerical value with a given threshold. If the calculated numerical value is higher than the threshold, the comprehensive determination section 104 may determine that a defect is present. In a case where the result indicating that a defect is absent is expressed by a numerical value of “ ⁇ 1” and the result indicating that a defect is present is expressed by a numerical value of “1”, the threshold may be set at “0”, which is an intermediate value between these numerical values. In this case, since 1.31>0, a final determination result given by the comprehensive determination section 104 indicates that a defect is present.
  • the comprehensive determination section 104 records, in the inspection result data 112 , the determination result obtained in S 24 . Then, in S 26 , a defect type determining process is carried out. Details of the defect type determining process will be described later with reference to FIGS. 14 and 15 .
  • the inspection image generating section 101 determines whether or not all the ultrasonic testing images 111 that are targets to be inspected have already been processed. If it is determined that there is an unprocessed ultrasonic testing image 111 (NO in S 27 ), the process returns to S 21 , where the inspection image generating section 101 reads out the unprocessed ultrasonic testing image 111 from the storage section 11 and generates an inspection image 111 A from the ultrasonic testing image 111 . Meanwhile, if it is determined that there is no unprocessed ultrasonic testing image 111 (YES in S 27 ), the process advances to S 28 .
  • the integrative detection section 108 integrates the defects detected by the comprehensive determination section 104 . Then, the integrative detection section 108 records the integration result in the inspection result data 112 .
  • a method for integrating the defects is as discussed with reference to FIG. 9 , and therefore is not described here again. If there are no defects to be integrated, the processes in S 28 and S 29 will not be carried out, and the process advances to S 30 .
  • the defect length calculating section 109 calculates a length of the defects integrated by the integrative detection section 108 .
  • the defect length calculating section 109 may calculate the length of the defect by multiplying (i) a length of a defect per ultrasonic testing image 111 by (ii) the number of defects integrated by the integrative detection section 108 . Then, the defect length calculating section 109 records the calculation result in the inspection result data 112 .
  • the thickness calculating section 107 calculates a wall thickness of the tube-to-tubesheet weld, and records the calculation result in the inspection result data 112 .
  • a method for calculating the wall thickness is as discussed above with reference to FIG. 10 , and therefore is not described here again.
  • the process shown in FIG. 13 is ended.
  • FIG. 14 is a flowchart illustrating an example of the defect type determining process.
  • the heat map generating section 105 generates a heat map with use of difference values (difference image) calculated when the determining section 102 A determines presence or absence of a defect.
  • the heat map generating section 105 carries out threshold processing on the heat map generated in S 41 .
  • the threshold processing is as discussed above with reference to FIG. 5 , and therefore is not described here again.
  • the defect type determining section 106 determines the type of the defect with use of the type decision model. Specifically, the defect type determining section 106 inputs, into the type decision model, the heat map having been subjected to the threshold processing in S 42 , and determines the type of the defect in accordance with a resulting output value. For example, in a case where the type decision model is constructed to output, for each type of defect, a numerical value indicating the likelihood that the defect corresponds to the type, the defect type determining section 106 may determine that the type of the defect is a type having a highest numerical value.
  • the defect type determining section 106 records, in the inspection result data 112 , the determination result obtained in S 43 . Then, the defect type determining process is ended.
  • the defect type determining section 106 may carry out, instead of the defect type determining process shown in FIG. 14 , a defect type determining process shown in FIG. 15 .
  • FIG. 15 is a flowchart illustrating an example of a defect type determining process that determines the type of a defect in accordance with the position of a defect area.
  • the defect type determining section 106 carries out the threshold processing on the difference values (difference image) calculated when the determining section 102 A determines presence or absence of a defect.
  • the threshold processing in S 51 is similar to the threshold processing in S 42 in FIG. 14 .
  • the defect type determining section 106 detects a defect area in accordance with the difference values after the threshold processing. A method for detecting the defect area is as discussed with reference to FIG. 7 , and therefore is not described here again.
  • the defect type determining section 106 determines the type of the defect in accordance with the position of the defect area identified in S 52 .
  • the defect type determining section 106 may determine the type of the defect in accordance with which of the areas AR 1 to AR 4 shown in FIG. 8 the defect area detected in S 52 is contained.
  • the defect type determining section 106 records, in the inspection result data 112 , the determination result obtained in S 53 . Then, the defect type determining process is ended.
  • the defect type determining section 106 may carry out both the defect type determining process shown in FIG. 14 and the defect type determining process shown in FIG. 15 .
  • the defect type determining section 106 may record the determination results obtained in both.
  • the defect type determining section 106 may integrate the two determination results together to make final determination of the type of the defect.
  • the defect type determining section 106 may calculate the reliability of the determination result given by the defect type determining process shown in FIG. 14 and the reliability of the determination result given by the defect type determining process shown in FIG. 15 , and may yield a final determination result regarding the type of the defect in accordance with the reliabilities thus calculated.
  • the reliabilities can be calculated in a similar manner to that for the reliability of the determination result given by the determining section 102 .
  • the foregoing embodiment has dealt with the example in which presence or absence of a defect in a tube-to-tubesheet weld is determined in accordance with an ultrasonic testing image 111 .
  • the determination matter may be any matter
  • the target data to be used for the determination may be any data selected in accordance with the determination matter.
  • the determination matter and the target data are not limited to those adopted in the foregoing embodiment.
  • the information processing device 1 is applicable to an inspection for determining presence or absence of a defect (which may also called “abnormal portion”) in an inspection target in radiographic testing (RT).
  • a defect which may also called “abnormal portion”
  • RT radiographic testing
  • an image related to an abnormal portion is detected from, in place of a radiograph, image data obtained with use of an electric device such as an imaging plate.
  • the determining section 102 A can determine presence or absence of a defect with use of a generative model
  • the determining section 102 C can determine presence or absence of an abnormal portion with use of a decision model
  • the determining section 102 B can determine presence or absence of an abnormal portion by numerical analysis in accordance with pixel values, size, and/or the like of the image captured in the image data.
  • the information processing device 1 is applicable to various kinds of nondestructive inspections that uses various data. Furthermore, the information processing device 1 is applicable to, in addition to the nondestructive inspections, detection of an object in a still image or a moving image and classification of the detected object, for example.
  • the foregoing embodiment has dealt with the example in which an output value obtained by inputting an inspection image into a reliability prediction model is used as a reliability.
  • the reliability may be any one, provided that it is derived from data used by the determining section 102 for determination.
  • the reliability prediction model for the determining section 102 B may be a model that accepts a binarized image as input data.
  • the reliability prediction model for the determining section 102 C may be a model that accepts an inspection image as input data.
  • the reliability prediction models for the determining sections 102 do not need to be constructed to accept completely the same input data.
  • the foregoing embodiment has dealt with the example in which the three determining sections 102 are employed.
  • the number of determining sections 102 may be two or four or more.
  • the determination methods of the three determining sections 102 differ from each other.
  • An entity that carries out each process described in each of the foregoing embodiments can be changed as appropriate.
  • the processes in S 21 (generation of an inspection image), S 23 (calculation with use of a reliability decision model), S 26 (determination of the type of a defect), S 28 (integration of defects), S 29 (calculation of a defect length), and S 30 (calculation of a wall thickness) in the flowchart shown in FIG. 13 may be carried out by another information processing device.
  • a part or all of the processes to be executed by the determining sections 102 A to 102 C may be executed by another information processing device. In these cases, the number of another information processing device(s) may be one or two or more.
  • the functions of the information processing device 1 can be realized by wide variety of system configurations. In a case where a system including a plurality of information processing devices is constructed, some of the plurality of information processing devices may be provided on cloud. That is, the functions of the information processing device 1 can also be realized by one information processing device or a plurality of information processing devices carrying out information processing online.
  • Control blocks of the information processing device 1 can be realized by a logic circuit (hardware) provided in an integrated circuit (IC chip) or the like or can be alternatively realized by software.
  • the information processing device 1 includes a computer that executes instructions of an information processing program that is software realizing the foregoing functions.
  • the computer for example, includes at least one processor and a computer-readable storage medium storing the information processing program.
  • An object of the present invention can be achieved by the processor of the computer reading and executing the information processing program stored in the storage medium.
  • Examples of the processor encompass a central processing unit (CPU).
  • the information processing device 1 may include, in addition to the processor such as CPU, a graphics processing unit (GPU). Use of GPU enables, e.g., high-speed computing involving use of the foregoing various models.
  • the storage medium can be a “non-transitory tangible medium” such as a tape, a disk, a card, a semiconductor memory, a programmable logic circuit as well as a read only memory (ROM) or the like.
  • the computer can further include a random access memory (RAM) in which the program is loaded.
  • the program may be made available to the computer via any transmission medium (such as a communication network and a broadcast wave) which enables transmission of the program.
  • any transmission medium such as a communication network and a broadcast wave
  • an aspect of the present invention can also be implemented by the program in the form of a computer data signal embedded in a carrier wave which is embodied by electronic transmission.
  • the present invention is not limited to the embodiments, but can be altered by a skilled person in the art within the scope of the claims.
  • the present invention also encompasses, in its technical scope, any embodiment derived by combining technical means disclosed in differing embodiments.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Theoretical Computer Science (AREA)
  • Biochemistry (AREA)
  • Immunology (AREA)
  • Pathology (AREA)
  • Analytical Chemistry (AREA)
  • Chemical & Material Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • Software Systems (AREA)
  • Multimedia (AREA)
  • Medical Informatics (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Quality & Reliability (AREA)
  • Investigating Or Analyzing Materials By The Use Of Ultrasonic Waves (AREA)
  • Image Analysis (AREA)

Abstract

A final determination result is derived in accordance with target data, in consideration of determination results given by determining sections. An information processing device includes: a reliability determining section that determines, in accordance with an inspection image, reliabilities of determination results given by determining sections each configured to determine a given determination matter in accordance with the inspection image; and a comprehensive determination section configured to determine the given determination matter with use of the determination results and the reliabilities.

Description

    TECHNICAL FIELD
  • The present invention relates to an information processing device and the like that determine a given determination matter (matter to be determined) in accordance with target data.
  • BACKGROUND ART
  • Patent Literature 1 indicated below discloses a technique that determines presence or absence of a defect and the type of the defect in accordance with an image of a semiconductor substrate. Specifically, the technique disclosed in Patent Literature 1 decides a final classification result in accordance with a total sum obtained by summing up values obtained by multiplying (i) respective classification results of classifiers by (ii) weights respectively determined in advance for the classifiers. With this, it is possible to expect improvement in classification accuracy as compared to that of a case in which only one classifier is used.
  • CITATION LIST Patent Literature
    • Patent Literature 1
    • Japanese Patent Application Publication, Tokukai, No. 2016-40650
    SUMMARY OF INVENTION Technical Problem
  • However, in a case where a defect is determined and classified in accordance with a plurality of images, the accuracy of classification by each classifier can vary depending on the image, since the images differ from each other in various points. Therefore, the weights determined in advance may not always be optimum. Using the weights that are not optimum can affect the accuracy in final determination.
  • For example, the following case may occur. That is, when two classifiers A and B are used, classification made by the classifier A is correct for a certain image but classification made by the classifier B is incorrect for the certain image, or vice versa for another image. In this case, if the weight on the classifier A is set to be greater than that of the classification B, a final classification result on the certain image is correct, but a final classification result on the another image is incorrect.
  • Such a problem is not limited to classification involving use of a plurality of classifiers, but is a common problem that can occur when a final determination result is derived in accordance with determination results given by a plurality of determining sections with regard to a given determination matter. Further, such a problem is not limited to determination involving use of an image, but is a common problem that can occur when determination is made in accordance with arbitrary target data.
  • An aspect of the present invention was made in view of the above-described problem, and has an object to realize an information processing device and the like capable of deriving a final determination result in appropriate consideration of determination results of determining sections in accordance with target data.
  • Solution to Problem
  • In order to attain the above object, an information processing device in accordance with an aspect of the present invention includes: a reliability determining section configured to carry out, for each of determining sections configured to determine a given determination matter in accordance with single target data, a process of determining a reliability, which is an indicator indicating a degree of certainty of a determination result given by the determining section, in accordance with the target data; and a comprehensive determination section configured to determine the given determination matter with use of the determination results and the reliabilities determined by the reliability determining section.
  • In order to attain the above object, a determination method in accordance with an aspect of the present invention is a determination method that is to be executed by one or more information processing devices, including the steps of: (i) carrying out, for each of determining sections each configured to determine a given determination matter in accordance with single target data, a process of determining a reliability, which is an indicator indicating a degree of certainty of a determination result given by the determining section, in accordance with the target data; and (ii) determining the given determination matter with use of the determination results and the reliabilities determined in the step (i).
  • Advantageous Effects of Invention
  • In accordance with an aspect of the present invention, it is possible to derive a final determination result in appropriate consideration of determination results given by determining sections in accordance with target data.
  • BRIEF DESCRIPTION OF DRAWINGS
  • FIG. 1 is a block diagram illustrating an example of a configuration of a main part of an information processing device in accordance with Embodiment 1 of the present invention.
  • FIG. 2 is a view illustrating an outline of an inspection system including the information processing device.
  • FIG. 3 is a view illustrating an outline of an inspection carried out by the information processing device.
  • FIG. 4 is a view illustrating an example of a configuration of a determining section that carries out determination with use of a generative model and an example of a method for determination, carried out by the determining section, of presence or absence of a defect.
  • FIG. 5 is a view illustrating an example in which a heat map is generated from an ultrasonic testing image and threshold processing is carried out on the heat map thus generated.
  • FIG. 6 is a view illustrating a relation between the position of a defect, an ultrasonic testing image, and a heat map.
  • FIG. 7 is a view illustrating a method for detecting a defect area.
  • FIG. 8 is a view illustrating an example of areas set for respective types of defect.
  • FIG. 9 is a view illustrating a method for integrating defects captured in a plurality of ultrasonic testing images to detect the defects as a single defect.
  • FIG. 10 is a view illustrating a method for calculating a thickness of a tube-to-tubesheet weld.
  • FIG. 11 is a view illustrating an example of output of an inspection result.
  • FIG. 12 is a view illustrating an example of a process of constructing various models used in an inspection and determining a threshold.
  • FIG. 13 is a view illustrating an example of an inspection method involving use of the information processing device.
  • FIG. 14 is a flowchart illustrating an example of a defect type determining process that determines the type of a defect with use of a type decision model.
  • FIG. 15 is a flowchart illustrating an example of a defect type determining process that determines the type of a defect in accordance with the position of a defect area.
  • DESCRIPTION OF EMBODIMENTS
  • Outline of System
  • The following description will discuss, with reference to FIG. 2 , an outline of an inspection system in accordance with an embodiment of the present invention. FIG. 2 is a view illustrating an outline of an inspection system 100. The inspection system 100 is a system that carries out an inspection to determine, in accordance with an image of an inspection target, presence or absence of a defect in the inspection target. The inspection system 100 includes an information processing device 1 and an ultrasonic testing device 7.
  • The description in the present embodiment will discuss an example in which the inspection system 100 carries out an inspection to determine presence or absence of a defect in a tube-to-tubesheet weld of a heat exchanger. Note that the tube-to-tubesheet weld refers to a part in which a plurality of metal tubes constituting the heat exchanger are welded to a metal tubesheet that bundles the tubes. The defect in the tube-to-tubesheet weld refers to a gap created inside the tube-to-tubesheet weld. Note that each of the tubes and the tubesheet may be made of a nonferrous metal such as aluminum or a resin. With the inspection system 100, it is also possible to carry out an inspection to determine presence or absence of a defect in a welded part (base welded part) between a tube support and a tube in boiler equipment used in a garbage incineration plant, for example. Needless to say, the part to be inspected is not limited to the welded part, and the inspection target is not limited to the heat exchanger.
  • An inspection is carried in the following manner. As shown in FIG. 2 , a probe having a contact medium applied thereto is inserted through a tube end. Then, the probe emits an ultrasonic wave so that the ultrasonic wave is propagated from an inner wall surface side of the tube toward the tube-to-tubesheet weld, and measures an echo of the ultrasonic wave. If such a defect as a gap in the tube-to-tubesheet weld occurs, an echo from the gap can be measured. In accordance with the echo, it is possible to detect the defect.
  • For example, the lower left part of FIG. 2 shows an enlarged view of an area around the probe. In the enlarged view, an ultrasonic wave indicated by the arrow L3 is propagated in a portion of the tube-to-tubesheet weld which portion has no gap. Thus, an echo of the ultrasonic wave indicated by the arrow L3 would not be measured. Meanwhile, an ultrasonic wave indicated by the arrow L2 is propagated toward a portion of the tube-to-tubesheet weld which portion has a gap. Thus, an echo of the ultrasonic wave reflected by the gap is measured.
  • Further, an ultrasonic wave is reflected also by the periphery of the tube-to-tubesheet weld, and therefore an echo of the ultrasonic wave propagated in the periphery is also measured. For example, since an ultrasonic wave indicated by the arrow L1 is propagated in a part closer to the tube end than the tube-to-tubesheet weld is, the ultrasonic wave does not hit the tube-to-tubesheet weld but is reflected by a tube surface of the part closer to the tube end than the tube-to-tubesheet weld is. Thus, due to the ultrasonic wave indicated by the arrow L1, an echo coming from the tube surface is measured. Meanwhile, an ultrasonic wave indicated by the arrow L4 is reflected by a tube surface of a part of the tube-to-tubesheet weld which part is closer to the far side of the tube. Thus, an echo of that ultrasonic wave is measured.
  • The tube-to-tubesheet weld surrounds the tube by 360 degrees. Thus, measurement is carried out repeatedly by circumferentially moving the probe by a certain angle (e.g., 1 degree). Then, data indicating the measurement result obtained with the probe is transmitted to the ultrasonic testing device 7. For example, the probe may be an array probe constituted by a plurality of array elements. In a case where the array probe is employed, the array probe may be disposed so that a direction of arrangement of the array elements coincides with a direction in which the tube extends. With this, it is possible to effectively inspect the tube-to-tubesheet weld whose width extends in the extending direction of the tube. Alternatively, the array probe may be a matrix array probe constituted by array elements arranged in rows and columns.
  • With use of the data indicated by the result of the measurement carried out by the probe, the ultrasonic testing device 7 generates an ultrasonic testing image that is an image of the echoes of the ultrasonic waves propagated in the tube and the tube-to-tubesheet weld. FIG. 2 illustrates an ultrasonic testing image 111, which is an example of the ultrasonic testing image generated by the ultrasonic testing device 7. Alternatively, the information processing device 1 may be configured to generate the ultrasonic testing image 111. In this case, the ultrasonic testing device 7 transmits, to the information processing device 1, the data indicating the measurement result obtained by the probe.
  • In the ultrasonic testing image 111, an intensity of a measured echo is presented as a pixel value of each pixel. An image area of the ultrasonic testing image 111 can be divided into a tube area ar1 corresponding to the tube, a welded area art corresponding to the tube-to-tubesheet weld, and peripheral echo areas ar3 and ar4 where echoes from peripheral parts of the tube-to-tubesheet weld appear.
  • As discussed above, the ultrasonic wave propagated from the probe in a direction indicated by the arrow L1 is reflected by the tube surface of the part closer to the tube end than the tube-to-tubesheet weld is. This ultrasonic wave is also reflected by the inner surface of the tube. These reflections occur repeatedly. Thus, repetition of echoes a1 to a4 appears in the peripheral echo area ar3, which extends along the arrow L1 in the ultrasonic testing image 111. The ultrasonic wave propagated from the probe in a direction indicated by the arrow L4 is repeatedly reflected by the outer surface and the inner surface of the tube. Thus, repetition of echoes a6 to a9 appears in the peripheral echo area ar4, which extends along the arrow L4 in the ultrasonic testing image 111. Each of these echoes, which appear in the peripheral echo areas ar3 and ar4, is also called “bottom echo”.
  • The ultrasonic wave propagated from the probe in a direction indicated by the arrow L3 is not reflected by anything. Thus, no echo appears in an area extending along the arrow L3 in the ultrasonic testing image 111. Meanwhile, the ultrasonic wave propagated from the probe in a direction indicated by the arrow L2 is reflected by the gap, i.e., the defect portion in the tube-to-tubesheet weld. Thus, an echo a5 appears in an area extending along the arrow L2 in the ultrasonic testing image 111.
  • The information processing device 1 analyzes such an ultrasonic testing image 111 to inspect whether or not the tube-to-tubesheet weld has a defect (details thereof will be described later). If the information processing device 1 determines that the tube-to-tubesheet weld has a defect, the information processing device 1 automatically determines the type of the defect, too.
  • Configuration of Information Processing Device
  • The following description will discuss a configuration of the information processing device 1 with reference to FIG. 1 . FIG. 1 is a block diagram illustrating an example of a configuration of a main part of the information processing device 1. As shown in FIG. 1 , the information processing device 1 includes a control section 10 which comprehensively controls the sections of the information processing device 1 and a storage section 11 in which various data used by the information processing device 1 is stored. The information processing device 1 further includes an input section 12 which accepts an input manipulation on the information processing device 1 and an output section 13 through which the information processing device 1 outputs data.
  • The control section 10 includes an inspection image generating section 101, a determining section 102A, a determining section 102B, a determining section 102C, a reliability determining section 103, a comprehensive determination section 104, a heat map generating section 105, a defect type determining section 106, a thickness calculating section 107, an integrative detection section 108, and a defect length calculating section 109. The storage section 11 has the ultrasonic testing image 111 and inspection result data 112 stored therein. In the description below, each of the determining sections 102A, 102B, and 102C will be referred to simply as a determining section 102, in a case where there is no need to distinguish the determining sections 102A, 102B, and 102B from each other.
  • The inspection image generating section 101 cuts an inspection target area from the ultrasonic testing image 111, so as to generate an inspection image used to determine presence or absence of a defect in the inspection target. A method for generating the inspection image will be described in detail later.
  • The determining section 102 determines a given determination matter in accordance with target data. In the example described in the present embodiment, the target data is an inspection image generated by the inspection image generating section 101, and the given determination matter is presence or absence of a welding defect in a tube-to-tubesheet weld in a heat exchanger captured in the inspection image. In the description below, the welding defect may simply be abbreviated as “defect”.
  • Note that the “defect” that is a determination target may be defined in advance in accordance with the purpose and/or the like of the inspection. For example, in a case of a quality inspection of a tube-to-tubesheet weld in a manufactured heat exchanger, it may be determined that a “defect” is present when the inspection image includes an echo caused by a gap inside the tube-to-tubesheet weld or a non-allowable recess on a surface of the tube-to-tubesheet weld. Such a recess is caused by burn-through, for example. The “presence or absence of a defect” can be reworded as presence or absence of a portion (abnormal portion) different from that in a normal product. In the field of nondestructive inspection, an abnormal portion detected with use of an ultrasonic waveform or an ultrasonic testing image is generally called “flaw”. The “flaw” is also encompassed in the “defect”. In addition, the “defect” further encompasses chipping and cracking.
  • Each of the determining sections 102A, 102B, and 102C determines presence or absence of a defect in accordance with an inspection image generated by the inspection image generating section 101. However, the determining methods of the determining sections 102A, 102B, and 102C differ from each other, as will be described later.
  • The determining section 102A (generative model determining section) determines presence or absence of a defect with use of a generated image generated by inputting the inspection image into a generative model. The determining section 102B (numerical analysis/determination section) analyzes pixel values in the inspection image to identify an inspection target portion in the inspection image, and determines presence or absence of a defect in accordance with pixel values in the inspection target portion thus identified. The determining section 102C determines presence or absence of a defect in accordance with an output value obtained by inputting the inspection image into a decision model. Details of determinations carried out by the determining sections 102A to 102C and various models to be used will be described later.
  • For each of the determination results given by the determining sections 102A to 102C, the reliability determining section 103 determines a reliability, which is an indicator indicating a degree of certainty of the determination result. The reliabilities are determined in accordance with the inspection image used by the determining sections 102A to 102C to derive the determination results (details thereof will be described later).
  • The comprehensive determination section 104 determines presence or absence of a defect with use of (i) the determination results given by the determining sections 102A to 102C and (ii) the reliabilities determined by the reliability determining section 103. With this, it is possible to obtain a determination result in appropriate consideration of the determination results given by the determining sections 102A to 102C with a reliability corresponding to the inspection image. Details of the determination method carried out by the comprehensive determination section 104 will be described later.
  • The heat map generating section 105 generates a heat map with use of data obtained in the determining process of the determining section 102A. The heat map is used by the defect type determining section 106 to determine the type of the defect. The heat map will be described in detail later.
  • For an inspection image determined as including a defect by the comprehensive determination section 104, the defect type determining section 106 determines the type of the defect included in the inspection image. As discussed above, the determination of the type is carried out with use of the heat map generated by the heat map generating section 105. A method for determining the type of the defect will be described later.
  • The thickness calculating section 107 calculates a wall thickness of the tube-to-tubesheet weld. The wall thickness calculated by the thickness calculating section 107 can be used as an indicator used to determine whether or not welding has been carried out appropriately. A method for calculating the wall thickness will be described later.
  • If the comprehensive determination section 104 determines that a defect is present in ultrasonic testing images 111 corresponding to parts of the inspection target which parts are adjacent to each other, the integrative detection section 108 detects, as a single defect, the defects captured in the ultrasonic testing images 111. The integration of the defect will be described in detail later.
  • The defect length calculating section 109 calculates a length of the defect integrated by the integrative detection section 108. A method for calculating the length of the defect will be described later.
  • As discussed above, the ultrasonic testing image 111 is an image of an echo of an ultrasonic wave propagated in the inspection target, and is generated by the ultrasonic testing device 7.
  • The inspection result data 112 refers to data indicating a result of a defect inspection carried out by the information processing device 1. Recorded in the inspection result data 112 is information indicating a result of determination of presence or absence of a defect, made by the comprehensive determination section 104, for the ultrasonic testing image 111 stored in the storage section 11. Further, for an ultrasonic testing image 111 determined as including a defect, the determination result as to the type of the defect made by the defect type determination section 106 is recorded in the inspection result data 112. Furthermore, information indicating the defect integrated by the integrative detection section 108, information indicating a length of the integrated defect calculated by the defect length calculating section 109, and information indicating a wall thickness of the tube-to-tubesheet weld calculated by the thickness calculating section 107 are recorded in the inspection result data 112.
  • Outline of Inspection
  • The following description will discuss, with reference to FIG. 3 , an outline of an inspection carried out by the information processing device 1. FIG. 3 is a view illustrating an outline of an inspection carried out by the information processing device 1. Note that FIG. 3 shows a process to be carried out after the ultrasonic testing image 111 generated by the ultrasonic testing device 7 is stored in the storage section 11 of the information processing device 1.
  • First, the inspection image generating section 101 extracts an inspection target area from the ultrasonic testing image 111 to generate an inspection image 111A. The extraction of the inspection target area may be carried out with use of an extraction model constructed by machine learning. The extraction model will be described with reference to FIG. 12 .
  • The inspection target area refers to an area sandwiched between two peripheral echo areas ar3 and ar4 in each of which an echo coming from the periphery of an inspection target portion of the inspection target appears repeatedly. As shown in FIG. 2 , in the periphery of the inspection target portion in the ultrasonic testing image 111, a given echo caused by the shape and/or the like of the peripheral part is repeatedly observed (echoes a1 to a4 and echoes a6 to a9). Thus, in accordance with the positions of the peripheral echo areas ar3 and ar4 in each of which such an echo repeatedly appears, it is possible to identify the area corresponding to the inspection target portion of the ultrasonic testing image 111. Note that it is not only the ultrasonic testing image 111 of the tube-to-tubesheet weld that a given echo appears in the periphery of an inspection target portion. Thus, the configuration that extracts, as the inspection target area, an area surrounded by the peripheral echo areas is applicable also to inspections on parts other than the tube-to-tubesheet weld.
  • Subsequently, the determining sections 102A, 102B, and 102C determine presence or absence of a defect in accordance with the inspection image 111A. The content of the determination will be described in detail later.
  • Then, the reliability determining section 103 determines reliabilities of the determination results given by the determining sections 102A, 102B, and 102C. Specifically, the reliability of the determination result given by the determining section 102A is determined in accordance with an output value obtained by inputting the inspection image 111A into a reliability prediction model for the determining section 102A. Similarly, the reliability of the determination result given by the determining section 102B is determined in accordance with an output value obtained by inputting the inspection image 111A into a reliability prediction model for the determining section 102B. The reliability of the determination result given by the determining section 102C is determined in accordance with an output value obtained by inputting the inspection image 111A into a reliability prediction model for the determining section 102C.
  • Then, the comprehensive determination section 104 comprehensively determines presence or absence of a defect in accordance with (i) the determination results given by the determining sections 102A, 102B, and 102C and (ii) the reliabilities of the determination results determined by the reliability determining section 103, and outputs a result of the comprehensive determination. This result is added to the inspection result data 112. The comprehensive determination section 104 may cause the output section 13 to output the result of the comprehensive determination.
  • In the comprehensive determination, the determination result given by the determining section 102 may be represented as a numerical value, and the reliability determined by the reliability determining section 103 may be used as a weight. For example, if the determining sections 102A, 102B, and 102C determine that a defect is present, “1” is output as the determination result. Meanwhile, if the determining sections 102A, 102B, and 102C determine that a defect is absent, “−1” is output as the determination result. The reliability determining section 103 outputs reliabilities within a numerical range from 0 to 1 (a value closer to 1 indicates a higher reliability).
  • In this case, the comprehensive determination section 104 may calculate a total value obtained by summing up values obtained by multiplying (i) the values (“1” or “−1”) output by the determining sections 102A, 102B, and 102C by (ii) the reliabilities output by the reliability determining section 103. Then, the comprehensive determination section 104 may determine presence or absence of a defect in accordance with whether or not the total value thus calculated is higher than a given threshold.
  • For example, assume that the threshold is set at “0”, which is intermediate between “1” indicating that a defect is present and “−1” indicating that a defect is absent. Assume also that the output values of the determining sections 102A, 102B, and 102C are respectively “1”, “−1”, and “1” and the reliabilities thereof are respectively “0.87”, “0.51”, and “0.95”.
  • In this case, the comprehensive determination section 104 carries out calculation as follows: 1×0.87+(−1)×0.51+1×0.95. The result of the calculation is 1.31, which is higher than “0”, i.e., the threshold. Thus, the result of the comprehensive determination made by the comprehensive determination section 104 indicates that a defect is present.
  • Correction of Reliabilities
  • It is empirical knowledge that an echo caused by a welding defect tends to appear at a location above a center position in the image area of the inspection image 111A. Thus, if the determining section 102 determines that a defect is present and an echo caused by a welding defect appears at a location above a center position in the image area of the inspection image 111A, it is considered that the determination result is highly likely to be correct.
  • Thus, in the image area of the inspection image 111A, an area where an echo caused by a welding defect often appears may be preliminarily set. Then, if an echo caused by a welding defect is detected in that area when the determining section 102 determines that a defect is present, the reliability determining section 103 may increase the reliability of the determination result. By correcting, in this manner, the reliability in consideration of the appearance tendency and/or characteristics of the defect, it is possible to make the reliability more appropriate.
  • For example, an area above a center of the image area of the inspection image 111A may be set as the above-described area. Then, if an echo caused by a welding defect is detected in that area, the reliability determining section 103 may add a given constant to the reliability calculated with use of the reliability prediction model. Meanwhile, if the echo caused by the welding defect is detected at a position outside the above-described area, the reliability determining section 103 may subtract a given constant from the reliability calculated with use of the reliability prediction model.
  • Note, however, that, addition of a constant to a reliability is preferably carried out so that the reliability after the addition does not exceed 1. Note also that subtraction of a constant from a reliability is preferably carried out such that the reliability after the subtraction is not below 0.
  • Needless to say, the method for correcting the reliability is not limited to the above example. Alternatively, for example, the image area of the inspection image 111A may be divided into much smaller areas. Then, a value to be added to the reliability may be set higher, as a position of an echo caused by a welding defect is in an area where a welding defect appears with a higher frequency. Further alternatively, for example, a value proportional to a distance between a position of a detected echo and a position where a welding defect appears with a highest frequency may be added to the reliability. Still further alternatively, a value inversely proportional to that distance may be subtracted from the reliability.
  • The reliability may be corrected in consideration of an element(s) other than the position. For example, even in a case where it is determined that a defect is present and an echo caused by a welding defect is detected, it is suspected that the determination result may be incorrect if a pixel value of the echo is low. Thus, as the pixel value of the echo caused by the welding defect is lower, the value of the reliability may be corrected to be lower. Meanwhile, as the pixel value of the echo caused by the welding defect is higher, the value of the reliability may be corrected to be higher.
  • The above-described correction of the reliability can be suitably applied to the determination results given by the determining sections 102A and 102B. This is due to the following reason. That is, firstly, in the process in which the determining section 102A makes determination, a difference image is calculated, and the difference image can be used to calculate the position and/or pixel value of the echo caused by the welding defect. Secondly, in the process in which the determining section 102B makes determination, the echo caused by the welding defect is detected, and the detection result given thereby can be used.
  • Determination by Determining Section 102A
  • As discussed above, the determining section 102A determines presence or absence of a defect with use of a generated image obtained by inputting an inspection image into a generative model. The generative model is constructed by machine learning that uses, as training data, an image of an inspection target in which a defect is absent, so that the generative model generates a new image having a similar feature to that of an image input into the generative model. Note that the “feature” is any information obtained from an image. For example, a distribution state, a variance, and the like of pixel values in the image are also included in the “feature”.
  • The generative model is constructed by machine learning that uses, as training data, an image of an inspection target in which a defect is absent. Thus, if an image of an inspection target in which a defect is absent is input into the generative model as the inspection image, it is highly likely that a new image having a similar feature to that of the inspection image is output as a generated image.
  • Meanwhile, if an image of an inspection target in which a defect is present is input into the generative model as the inspection image, it is highly likely that a resulting generated image has a different feature from that of the inspection image, regardless of the position, shape, and size of the defect captured in the inspection image.
  • As discussed above, (i) the generated image generated from the inspection image in which a defect is captured and (ii) the generated image generated from the inspection image in which no defect is captured differ from each other in that one does not properly restore the target image input into the generative model and the other properly restores the target image input into the generative model.
  • Thus, with the information processing device 1 that carries out comprehensive determination in consideration of the determination result given by the determining section 102A that determines presence or absence of a defect with use of the generated image generated by the generative model, it is possible to determine, with high accuracy, presence or absence of a defect having irregular position, size, shape, and/or the like.
  • The following description will discuss, with reference to FIG. 4 , details of determination made by the determining section 102A. FIG. 4 is a view illustrating an example of a configuration of the determining section 102A and an example of a method for determination, carried out by the determining section 102A, of presence or absence of a defect. As shown in FIG. 4 , the determining section 102A includes an inspection image obtaining section 1021, a restored image generating section 1022, and a defect presence/absence determination section 1023.
  • The inspection image obtaining section 1021 obtains an inspection image. As discussed above, the information processing device 1 includes the inspection image generating section 101. Thus, the inspection image obtaining section 1021 obtains the inspection image generated by the inspection image generating section 101. Note that the inspection image may be generated by another device. In this case, the inspection image obtaining section 1021 obtains the inspection image generated by the another device.
  • The restored image generating section 1022 inputs, into the generative model, the inspection image obtained by the inspection image obtaining section 1021, so as to generate a new image having a similar feature to that of the inspection image thus input. Hereinafter, the image generated by the restored image generating section 1022 is called “restored image”. The generative model used to generate the restored image is also called “autoencoder”, and is constructed by machine learning that uses, as training data, an image of an inspection target in which a defect is absent (details thereof will be described later).
  • The defect presence/absence determining section 1023 determines whether or not the inspection target has a defect with use of the restored image generated by the restored image generating section 1022. Specifically, the defect presence/absence determining section 1023 determines that the inspection target has a defect, if a variance of pixel-by-pixel difference values between the inspection image and the restored image exceeds a given threshold.
  • In the method for determining, by the determining section 102A configured as above, presence or absence of a defect, the inspection image obtaining section 1021 first obtains the inspection image 111A. Then, the inspection image obtaining section 1021 transmits the obtained inspection image 111A to the restored image generating section 1022. As discussed above, the inspection image 111A is an image generated by the inspection image generating section 101 from the ultrasonic testing image 111.
  • Then, the restored image generating section 1022 inputs the inspection image 111A into the generative model, so as to generate a restored image 111B in accordance with a resulting output value. A method for generating the generative model will be described later.
  • Then, the inspection image obtaining section 1021 removes the peripheral echo areas from the inspection image 111A to generate a removed image 111C, and removes the peripheral echo areas from the restored image 111B to generate a removed image (restored) 111D. Note that the positions and sizes of the peripheral echo areas captured in the inspection image 111A are substantially the same, provided that the same inspection target is captured. Thus, the inspection image obtaining section 1021 may remove, as a peripheral echo area, a given range in the inspection image 111A. The inspection image obtaining section 1021 may analyze the inspection image 111A to detect the peripheral echo areas, and may remove the peripheral echo areas in accordance with a detection result.
  • As a result of removing the peripheral echo areas in the above-described manner, the defect presence/absence determining section 1023 determines presence or absence of a defect, with respect to a remaining image area obtained by removing the peripheral echo areas from the image area of the restored image 111B. Consequently, it is possible to determine presence or absence of a defect, without being affected by an echo coming from the periphery. This makes it possible to improve the accuracy in determination of presence or absence of a defect.
  • Next, the defect presence/absence determining section 1023 determines presence or absence of a defect. Specifically, the defect presence/absence determining section 1023 first calculates, in pixels, a difference between the removed image 111C and the removed image (restored) 111D. Next, the defect presence/absence determining section 1023 calculates a variance of the difference thus obtained. Then, the defect presence/absence determining section 1023 determines presence or absence of a defect in accordance with whether or not the value of the variance thus calculated exceeds a given threshold.
  • A difference value calculated for a pixel in which an echo caused by a defect appears is higher than difference values calculated for the other pixels. Thus, a variance of difference values calculated for a removed image 111C and a removed image (restored) 111D generated from an inspection image 111A where an echo caused by a defect is captured is large.
  • Meanwhile, a variance of difference values calculated for a removed image 111C and a removed image (restored) 111D generated from an inspection image 111A where an echo caused by a defect is not captured is relatively small. Meanwhile, in the case where the echo caused by the defect is not captured, a part having somewhat high pixel values due to the effects of noises and/or the like can occur, but a part having extremely high pixel values occurs with low probability.
  • The increase in variance of difference values is a phenomenon characteristic to a case where the inspection target has a defect. Thus, with the defect presence/absence determining section 1023 configured to determine that a defect is present if a variance of difference values exceeds a given threshold, it is possible to appropriately determine presence or absence of a defect.
  • For an inspection image 111A determined as including a defect, the defect type determining section 106 determines the type of the defect in accordance with the values of the difference in the pixels calculated by the defect presence/absence determining section 1023. Since the values of the difference in the pixels indicate the difference between the removed image 111C and the removed image (restored) 111D, these values are also called “difference image”.
  • Note that a timing to remove the peripheral echo areas is not limited to the above-described example. Alternatively, for example, a difference image between the inspection image 111A and the restored image 111B may be generated, and the peripheral echo areas may be removed from the difference image.
  • Determination by Determining Section 102B
  • As discussed above, the determining section 102B analyzes pixel values in the inspection image, which is an image of the inspection target, to identify an inspection target portion in the inspection image, and determines presence or absence of a defect in accordance with pixel values in the inspection target portion thus identified.
  • In a conventional inspection involving use of an image, an inspector visually carries out a process of identifying an inspection target portion in an image and checking, in the identified portion, for a defect such as a damage and/or a gap that should not exist from a design standpoint. Such a visual inspection is requested to be automated, from the viewpoints of reduction of labor, achievement of stable accuracy, and/or the like.
  • The determining section 102B analyzes the pixel values in the image to identify an inspection target portion in the image, and determines presence or absence of a defect in accordance with pixel values in the inspection target portion thus identified. Thus, it is possible to automate the above-described visual inspection. Further, the information processing device 1 makes determination by comprehensively considering the determination result given by the determining section 102B and the determination result(s) given by other determining section(s) 102. Therefore, it is possible to determine, with high accuracy, presence or absence of a defect.
  • The following will give a more detailed description of the content of a process (numerical analysis) to be executed by the determining section 102B. First, in the inspection image, the determining section 102B identifies, as the inspection target portion, an area sandwiched between two peripheral echo areas (peripheral echo areas ar3 and ar4 in the example shown in FIG. 2 ) in each of which an echo coming from the periphery of the inspection target portion appears repeatedly. Then, the determining section 102B determines presence or absence of a defect in accordance with whether or not the identified inspection target portion includes an area (also called “defect area”) constituted by pixel values not less than a threshold.
  • In order to detect the peripheral echo areas and the defect area, the determining section 102B may first binarize the inspection image 111A with use of a given threshold to generate a binarized image. Then, the determining section 102B detects the peripheral echo areas from the binarized image. For example, the inspection image 111A shown in FIG. 3 includes echoes a1, a2, a6, and a7. By binarizing the inspection image 111A with use of such a threshold that can divide these echoes and noise components from each other, the determining section 102B can detect these echoes in the binarized image. Then, the determining section 102B can detect edges of these echoes thus detected, and can identify, as the inspection target portion, an area surrounded by these edges.
  • To be more specific, the determining section 102B identifies a right edge of the echo a1 or a2 as a left edge of the inspection target portion, and identifies a left edge of the echo a6 or a7 as a right edge of the inspection target portion. These edges constitute boundaries between (i) the peripheral echo areas ar3 and ar4 and (ii) the inspection target portion. Similarly, the determining section 102B identifies an upper edge of the echo a1 or a6 as an upper edge of the inspection target portion, and identifies a lower edge of the echo a2 or a7 as a lower edge of the inspection target portion.
  • Note that, as in the ultrasonic testing image 111 shown in FIG. 2 , an echo caused by a defect may appear at a location above the echoes a1 and a6. Thus, the determining section 102B may set the upper edge of the inspection target portion at a location above the upper edge of the echo a1 or a6.
  • Further, the determining section 102B can analyze the inspection target portion identified in the binarized image to determine whether or not the echo caused by the defect is captured therein. For example, in a case where the inspection target portion includes a continuous area constituted by a given number or more of pixels, the determining section 102B may determine that the echo caused by the defect is captured at a location where the continuous area exists.
  • Note that the above-described numerical analysis is an example, and the content of the numerical analysis is not limited to the above-described example. For example, in a case where there exists a significant difference between (i) a variance of pixel values in the inspection target portion having a defect and (ii) a variance of pixel values in the inspection target portion not having a defect, the determining section 102B may determine presence or absence of a defect in accordance with the value of the variance.
  • Determination by Determining Section 102C
  • As discussed above, the determining section 102C determines presence or absence of a defect in accordance with an output value obtained by inputting the inspection image into a decision model. The decision model is constructed by, e.g., carrying out machine learning with use of (i) training data generated by using an ultrasonic testing image 111 of an inspection target in which a defect is present and (ii) training data generated by using an ultrasonic testing image 111 of an inspection target in which a defect is absent.
  • The decision model can be constructed by any learning model suitable for image classification. For example, the decision model may be constructed by, e.g., convolutional neural network having excellent image classification accuracy.
  • Heat Map and Threshold Processing
  • As discussed above, the heat map is used to determine the type of the defect. Here, the following description will discuss, with reference to FIG. 5 , the heat map generated by the heat map generating section 105 and the threshold processing carried out on the heat map thus generated. FIG. 5 is a view illustrating an example in which a heat map is generated from an ultrasonic testing image and threshold processing is carried out on the heat map thus generated. To be more specific, the upper part of FIG. 5 illustrates an example of an ultrasonic testing image 111-a of a portion of a tube-to-tubesheet weld in which a defect is present, whereas the lower part of FIG. 5 illustrates an example of an ultrasonic testing image 111-b of a portion of the tube-to-tubesheet weld in which a defect is absent.
  • As discussed with reference to FIG. 4 , in the process in which the determining section 102A carries out determination, an inspection image 111A is generated from the ultrasonic testing image 111, and a restored image 111B is generated from the inspection image 111A. Then, a removed image 111C is generated from the inspection image 111A, and a removed image (restored) 111D is generated from the restored image 111B.
  • In the example shown in FIG. 5 , a removed image 111C-a and a removed image (restored) 111D-a are generated from the ultrasonic testing image 111-a. A difference image is generated from the removed image 111C-a and the removed image (restored) 111D-a. The heat map generating section 105 generates a heat map in which pixels in the difference image are expressed by colors or gradations corresponding to the pixel values.
  • FIG. 5 shows a heat map 111E-a in which pixel values, from the lower limit to the higher limit, are expressed by gradations of colors from black to white. As indicated by the hollow arrow in the heat map 111E-a, an area corresponding to the defect (i.e., an area in which pixels having high pixel values are collected) is an area in which nearly white pixels are collected. Thus, in the heat map 111E-a, the area corresponding to the defect can be easily visually recognized.
  • Note, however, that the heat map 111E-a also has an area in which pixel values have become high due to noises and/or the like. Thus, it is preferable that the heat map generating section 105 carry out the threshold processing on the heat map thus generated, so as to correct the pixel values in the area in which the pixel values have become high due to noises and/or the like. For example, the heat map generating section 105 may set, at zero (black), a pixel value(s) in the heat map 111E-a which pixel value(s) is/are not more than a given threshold. Consequently, a heat map 111F-a from which a noise component(s) has/have been removed is generated. With the heat map 111F-a, the area corresponding to the defect can be more clearly recognized.
  • This is also true of an ultrasonic testing image 111-b of a portion in which a defect is absent. A removed image 111C-b and a removed image (restored) 111D-b are generated from the ultrasonic testing image 111-b, and a difference image is generated from the removed image 111C-b and the removed image (restored) 111D-b. Then, the heat map generating section 105 generates a heat map 111E-b of the difference image, and carries out the threshold processing on the heat map 111E-b to generate a heat map 111F-b. It is understood that comparison of the heat map 111F-a with the heat map 111F-b makes it possible to clearly determine presence or absence of a defect. It is also understood that the position of the defect can be clearly identified in the heat map 111F-a.
  • Type of Defect, Ultrasonic Testing Image, and Heat Map
  • As the defect in the tube-to-tubesheet weld, incomplete penetration in the first layer, incomplete fusion between welding passes, undercut, and a blowhole are known, for example. The incomplete penetration in the first layer refers to creation of a gap due to incomplete welding occurred in the vicinity of the tubesheet. The incomplete fusion between the welding passes refers to creation of a gap due to poor welding occurred while carrying out welding plural times. The undercut refers to such a defect that an end of a weld bead is hollowed out in the form of a notch. The blowhole refers to a spherical void formed in the welding metal.
  • The positions where these defects occur differ from each other. Thus, in accordance with the position in the ultrasonic testing image 111 at which position an echo caused by the defect appears, it is possible to determine the type of the defect. Similarly, it is also possible to determine the type of the defect in accordance with the position of the defect area in the heat map (preferably, the one after the threshold processing) generated from the ultrasonic testing image 111. As discussed above, the defect area is an area where an echo caused by the defect appears, and therefore pixel values therein are higher than those in the other areas.
  • The following description will discuss, with reference to FIG. 6 , determination of the type of the defect in accordance with the position of the defect area. FIG. 6 is a view illustrating a relation between the position of the defect, the ultrasonic testing image, and the heat map. The view at the left end in the first row in FIG. 6 shows a cross section of a tube-to-tubesheet weld in which incomplete penetration in the first layer occurs. The left side in FIG. 6 is the tube end side, whereas the right side in FIG. 6 is the far side of the tube. That is, the tube extends in a left-right direction in FIG. 6 . The tubesheet is located on a lower side relative to the tube wall. A scale is put on an inner wall surface (inner surface) of the tube so as to indicate a width of the tube-to-tubesheet weld.
  • In the view at the left end of the first row in FIG. 6 , an area indicated by the broken line is a welded area of the tubesheet obtained by welding, an inverted triangle area on the left side of the welded area is an area made of a welding metal, and an area obtained by combining these areas is the tube-to-tubesheet weld. In the circled portion of the tube-to-tubesheet weld, a gap is created. This gap is positioned at a location that is in the vicinity of the surface of the tube and that is closer to the end of the tube-to-tubesheet weld which end is closer to the far side of the tube.
  • As shown in the view at the center of the first row shown in FIG. 6 , an echo caused by the gap appears in an ultrasonic testing image 111-c of the portion including the gap. As shown in the view at the right end of the first row in FIG. 6 , as indicated by the hollow arrow shown in this view, an area corresponding to the gap appears also in a heat map 111F-c generated from the ultrasonic testing image 111-c.
  • The view at the left end in the second row in FIG. 6 shows a cross section of a tube-to-tubesheet weld in which incomplete fusion between welding passes occurs. In the circled portion, a gap is created. This gap is positioned at a location that is in the vicinity of the surface of the tube and that is near a center part in a thickness direction of the tube-to-tubesheet weld.
  • As shown in the view at the center of the second row shown in FIG. 6 , an echo caused by the gap appears in an ultrasonic testing image 111-d of the portion including the gap. As shown in the view at the right end of the second row in FIG. 6 , as indicated by the hollow arrow shown in this view, an area corresponding to the gap appears also in a heat map 111F-d generated from the ultrasonic testing image 111-d. This area is located more leftward than that in the heat map 111F-c on the first row.
  • The view at the left end in the third row in FIG. 6 shows a tube-to-tubesheet weld in which undercut occurs, viewed from the tube end side. In the circled portion, a gap is created. This gap is positioned at a location that is in the vicinity of the surface of the tube and that is near an end of the tube-to-tubesheet weld which end is closer to the tube end side.
  • As shown in the view at the center of the third row shown in FIG. 6 , an echo caused by the gap appears in an ultrasonic testing image 111-e of the portion including the gap. As shown in the view at the right end of the third row in FIG. 6 , as indicated by the hollow arrow shown in this view, an area corresponding to the gap appears also in a heat map 111F-e generated from the ultrasonic testing image 111-e. This area is located more leftward than that in the heat map 111F-d on the second row.
  • The view at the left end in the fourth row in FIG. 6 shows a cross section of a tube-to-tubesheet weld in which a blowhole is created. In the circled portion, a gap is created. This gap is positioned at a location that is closer to the inside of the tube-to-tubesheet weld rather than to the surface of the tube, and the position of this gap in a left-right direction is in the vicinity of a center of the tube-to-tubesheet weld in its width direction.
  • As shown in the view at the center of the fourth row shown in FIG. 6 , an echo caused by the gap appears in an ultrasonic testing image 111-f of the portion including the gap. As shown in the view at the right end of the fourth row in FIG. 6 , as indicated by the hollow arrow shown in this view, an area corresponding to the gap appears also in a heat map 111F-f generated from the ultrasonic testing image 111-f. The position of this area in the left-right direction is close to that in the heat map 111F-d in the second row. However, the position of this area in an up-down direction is more downward than that in the heat map 111F-d in the second row.
  • As discussed above, there is a correlation between the type of the defect and the appearance of the heat map 111F. Thus, on the basis of the correlation, it is possible to construct a type decision model used to determine the type of the defect from the heat map 111F. Such a type decision model can be constructed by machine learning that uses, as training data, a heat map of a difference image generated from an inspection image of an inspection target having a defect of a known type. Then, the defect type determining section 106 can determine the type of the defect in accordance with an output value obtained by inputting, into such a decision model, a heat map generated by the heat map generating section 105.
  • As discussed above, a heat map representing, by colors or gradations, pixel values of pixels constituting a difference image reflects a difference in the type of the defect captured in an inspection image from which the difference image is obtained. Thus, with the above configuration, it is possible to automatically determine the type of the defect in an appropriate manner.
  • For example, a large number of heat maps like the heat map 111F-c shown in FIG. 6 , which is generated from the ultrasonic testing image 111 of the portion in which incomplete penetration in the first layer occurs, may be prepared and used as training data. With this, it is possible to construct a type decision model that outputs a probability that the type of a defect is incomplete penetration in the first layer. Similarly, carrying out machine learning by using, as training data, a heat map generated from an ultrasonic testing image 111 of a portion in which another type of defect occurs makes it possible to construct a type decision model that outputs probabilities of various types of defects.
  • Thus, the defect type determining section 106 can determine the type of the defect in accordance with the output value obtained by inputting a heat map into such a type decision model. For example, the defect type determining section 106 may determine that a defect of the type corresponding to, among the probability values corresponding to various types of defects output from the type decision model, a highest probability value occurs.
  • Other Examples of Method for Determining Type of Defect
  • The following description will discuss, with reference to FIGS. 7 and 8 , other examples of the method for determining the type of the defect. In the below-described determination method, the defect type determining section 106 detects a defect area from a difference image, and determines the type of the defect related to the defect area, in accordance with the position in an image area of the difference image at which position the defect area is detected.
  • The following description will discuss, with reference to FIG. 7 , a method for detecting a defect area. FIG. 7 is a view illustrating the method for detecting the defect area. FIG. 7 illustrates an example in which a defect area is detected with use of a heat map. However, the generation of the heat map is not essential, as discussed below.
  • FIG. 7 shows (i) a heat map 111E generated from an ultrasonic testing image 111 of an inspection target in which a defect is present and (ii) a heat map 111F obtained by carrying out threshold processing on the heat map 111E. FIG. 7 also shows an enlarged view of the upper left part of the heat map 111F. In this enlarged view, the pixels in the heat map 111F indicate their pixel values.
  • In order to detect the defect area, the defect type determining section 106 first detects, in the heat map 111F, a pixel having a highest pixel value. In the example shown in FIG. 7 , the highest pixel value is 104. Thus, the pixel having this pixel value is detected. Next, the defect type determining section 106 detects a pixel(s) being adjacent to the detected pixel and having a pixel value(s) not less than a given threshold.
  • The defect type determining section 106 carries out such a process repeatedly until no adjacent pixel having a pixel value not less than the threshold is detected. Consequently, the defect type determining section 106 can detect, as a defect area, a continuous area constituted by the pixels each having a pixel value not less than the given threshold. The defect type determining section 106 may detect, as the defect area, a quadrangular area ar5 including the defect area detected in the above-described manner.
  • The above-discussed process can be carried out, if there is data indicating pixel-by-pixel difference values, i.e., a difference image between the inspection image 111A and the restored image 111B. In other words, by repeating the process of (i) detecting a pixel having a highest pixel value in the difference image and (ii) detecting a pixel(s) being adjacent to that pixel and having a pixel value(s) not less than the given threshold, it is possible to detect a defect area. Therefore, as discussed above, it is not essential to generate the heat map 111E or the heat map 111F in order to detect the defect area.
  • As discussed above, the defect type determining section 106 detects, as a defect area, an area constituted by a plurality of pixels which are included in a difference image and which have pixel values not less than a threshold. In the difference image, the pixel values of the pixels in the defect area are higher than the pixel values in the other areas. Thus, with the above configuration, it is possible to automatically detect an appropriate defect area.
  • As discussed with reference to FIG. 6 , as a defect in a welded portion, various types of defects are known, such as incomplete penetration in the first layer and incomplete fusion between welding passes. The difference in the defect type appears as a difference in position in an ultrasonic testing image. By utilizing this, the defect type determining section 106 determines the type of a defect related to a defect area, in accordance with the position in the image area of the difference image at which position the defect area is detected. With this, it is possible to automatically determine the type of the defect.
  • For example, if areas corresponding to various types of defects are preliminarily set in a difference image, the defect type determining section 106 can determine the type of the defect in accordance with which of the areas the defect area detected in the above-described manner is contained.
  • FIG. 8 is a view illustrating an example of areas set for respective types of defect. In the example shown in FIG. 8 , an area AR1 corresponding to undercut is set at the upper left corner, an area AR2 corresponding to incomplete fusion between welding passes is set at the upper center, and an area AR3 corresponding to incomplete penetration in the first layer is set at the upper right corner in the heat map 111F. In addition, an area AR4 corresponding to a blowhole is set at a location slightly above the center. These areas may be set in advance in accordance with, e.g., analysis on a difference image and/or a heat map that is/are based on inspection images of portions in which various defects are present. In the example shown in FIG. 8 , the defect area indicated by the hollow arrow is detected in the area AR3. Thus, the defect type determining section 106 determines that this defect is caused by incomplete penetration in the first layer.
  • In the example shown in FIG. 8 , a part of the area AR4, which corresponds to a blowhole, overlaps parts of the areas AR1 to AR3. In this manner, an area used to determine the type of defect may be set so as to partially overlap another area.
  • In this case, when a defect area is detected in an area where a plurality of areas overlap each other, the defect type determining section 106 may determine, as a determination result regarding the type of the defect, all the types corresponding to the plurality of areas. For example, when a defect area is detected in an area where the areas AR1 and AR4 overlap each other, the defect type determining section 106 may output, as a determination result, both undercut and a blowhole.
  • Further, the defect type determining section 106 may narrow down the determination results of the type of the defect in accordance with whether or not a condition(s) specific to each type of defect is/are satisfied. For example, in a case of a defect that is characterized by shape, a condition related to the shape may be set. Meanwhile, in a case of a defect that is characterized by size, a condition related to the size may be set.
  • Specifically, for example, a blowhole is such a defect that causes a spherical cavity, a diameter of which is generally not more than 2 mm. Thus, in a case where a single ultrasonic testing image 111 covers a range of approximately 1 mm of the inspection target in width, a single blowhole will fit within two or three ultrasonic testing images 111 or so. Therefore, if a defect is detected consecutively in ultrasonic testing images 111 respectively corresponding to adjacent parts of the inspection target and the number of ultrasonic testing images 111 is not more than three, the defect may possibly be a blowhole. Meanwhile, if the number of ultrasonic testing images 111 in which a defect is detected consecutively is not less than four, it is highly likely that the defect is not a blowhole.
  • Thus, in a case where a defect area is detected in an area where the area AR4 and another area overlap each other, the defect type determining section 106 may determine that the type of the defect is a blowhole, if the condition that the number of ultrasonic testing images 111 in which the defect is consecutively detected is not more than a threshold (e.g., 3) is satisfied.
  • For example, assume that, in the example shown in FIG. 8 , a defect area is detected in an area where the areas AR4 and AR2 overlap each other. In this case, if the number of ultrasonic testing images 111 in which the defect is consecutively detected is not more than the threshold, the defect type determining section 106 may determine that the type of the defect is a blowhole. Meanwhile, if the number of ultrasonic testing images 111 in which the defect is consecutively detected exceeds the threshold, the defect type determining section 106 may determine that the type of the defect is incomplete fusion between welding passes.
  • As discussed above, a blowhole has a spherical shape. Thus, if a single blowhole is detected across a plurality of ultrasonic testing images 111, the peak values of the echoes caused by the blowhole observed in the ultrasonic testing images 111 often differ from each other. Such a difference in peak value appears as a difference in pixel value in the ultrasonic testing images 111. For example, suppose that a single blowhole is detected across three ultrasonic testing images 111. In this case, if a peak value of an echo caused by the blowhole observed in the intermediate one of the three ultrasonic testing images 111 is 50%, a peak value of an echo caused by the blowhole in each of the ultrasonic testing images 111 before and after the intermediate one is 30%, which is lower than that of the intermediate one.
  • Thus, in a case where a defect area is detected in an area where the area AR4 and another region overlap each other, the defect type determining section 106 may determine that the type of the defect is a blowhole, if the condition that there is a difference between the pixel values in the defect areas in the ultrasonic testing images 111 in which the defect is consecutively detected is satisfied. For example, the defect type determining section 106 may calculate average values of the pixel values in the pixels included in the defect areas in the ultrasonic testing images 111. Then, if a difference between the average values is not less than a threshold, the defect type determining section 106 may determine that there is a difference.
  • Note that the defect type determining section 106 may carry out either or both of (i) determination involving use of the type decision model and (ii) determination in accordance with which of the areas the defect area is contained. Carrying out both of the determinations makes it possible to enhance the accuracy in the determination result regarding the type.
  • Integration of Defects
  • The tube-to-tubesheet weld surrounds the tube by 360 degrees. Thus, as discussed above, ultrasonic testing images 111 of various parts of the tube-to-tubesheet weld are generated by circumferentially moving, in the tube, the probe by a given degree. Then, in accordance with the ultrasonic testing images 111, a defect is detected. In such a case, there may be a case where a single continuous defect is captured in a plurality of ultrasonic testing images and accordingly the defect is determined as two or more defects, although the entity of the defect is a single defect.
  • In order to deal with this, the integrative detection section 108 integrates the defects captured in the plurality of ultrasonic testing images 111 so as to detect the defects as a single defect. To be more specific, if the comprehensive determination section 104 determines that a defect is present in a plurality of ultrasonic testing images 111 corresponding to parts of the tube-to-tubesheet weld which parts are adjacent to each other, the integrative detection section 108 detects, as a single defect, the defects captured in the plurality of ultrasonic testing images 111. This makes it possible to carry out detection appropriate to the entity of the defect.
  • The following description will discuss a method for integration of defects with reference to FIG. 9 . FIG. 9 is a view illustrating a method for integrating defects captured in a plurality of ultrasonic testing images 111 to detect the defects as a single defect. The upper left part of FIG. 9 illustrates a transverse cross section of a tube and a tube-to-tubesheet weld. The lower left part of FIG. 9 illustrates a longitudinal cross section of the tube, the tube-to-tubesheet weld, and a tubesheet.
  • In the example shown in FIG. 9 , a welding defect occurs in a wide range along an outer wall surface of the tube. When measurement of an echo is carried out while circumferentially moving the probe by a given degree along an inner wall surface of the tube, measurement results obtained in a range in which a welding defect occurs reflect echoes from the welding defect. Consequently, as shown in the right part of FIG. 9 , echoes caused by the welding defect appear in ultrasonic testing images 111 g to 111 i generated in accordance with the measurement results. Thus, in the determination of presence or absence of a defect from the ultrasonic testing images 111 g to 111 i, the comprehensive determination section 104 determines that a defect is present.
  • Here, the ultrasonic testing images 111 g to 111 i respectively correspond to parts of the tube-to-tubesheet weld which parts are adjacent to one another. Thus, the integrative detection section 108 detects, as a single defect, the defects captured in the ultrasonic testing images 111 g to 111 i that are determined as including a defect by the comprehensive determination section 104.
  • Note that the integrative detection section 108 may integrate the defects, provided that the positions of the defects detected in the ultrasonic testing images 111 g to 111 i are the same or close to each other. As discussed above, the position of the defect varies depending on the type. Thus, the integrative detection section 108 may integrate these defects, provided that the defects of the same type are detected in the ultrasonic testing images 111 g to 111 i. With these configurations, it is possible to enhance the accuracy in defect integration.
  • The defect length calculating section 109 calculates a length of the defect integrated by the above-described processing. For example, the defect length calculating section 109 may calculate the length of the defect by multiplying (i) a length of a defect per ultrasonic testing image 111 by (ii) the number of defects integrated by the integrative detection section 108.
  • For example, assume the following situation. That is, for a tube-to-tubesheet weld formed to surround a tube by 360 degrees, echo measurement is carried out 360 times by circumferentially moving the probe by 1 degree along an inner wall surface of the tube around a center axis of the tube, so that 360 ultrasonic testing images 111 are generated. In such a situation, a defect captured in a single ultrasonic testing image 111 has a length of approximately (outer diameter of tube)×π×1/360. Thus, in a case where three ultrasonic testing images 111 g to 111 i are integrated as in the manner shown in FIG. 9 , the defect length calculating section 109 may calculate the length of the defect as follows: (outer diameter of tube)×π×3×1/360. Note that π denotes pi.
  • Calculation of Thickness of Tube-to-Tubesheet Weld
  • The following description will discuss, with reference to FIG. 10 , a method for calculating a thickness (wall thickness) of a tube-to-tubesheet weld. FIG. 10 is a view illustrating the method for calculating the thickness of the tube-to-tubesheet weld. The lower part of FIG. 10 illustrates a longitudinal cross section of the tube-to-tubesheet weld, and the upper part of FIG. 10 illustrates an ultrasonic testing image 111 of the tube-to-tubesheet weld.
  • A thickness of the tube-to-tubesheet weld shown in the lower part of FIG. 10 , including a part penetrated into the tubesheet, is X. As discussed with reference to FIG. 2 , the area of the ultrasonic testing image 111 in which area the tube-to-tubesheet weld is captured is an area sandwiched between two peripheral echo areas ar3 and ar4 in each of which an echo coming from the periphery of the peripheral echo area appears repeatedly. Thus, the thickness X of the tube-to-tubesheet weld can be calculated on the basis of a distance Xi between the peripheral echo areas ar3 and ar4.
  • The distance Xi may be calculated by the thickness calculating section 107 analyzing the ultrasonic testing image 111. Alternatively, since the analysis of the ultrasonic testing image 111 is carried out by the determining section 102B, it is preferable to use the result of the analysis.
  • To be more specific, as discussed with reference to FIG. 3 , the determining section 102B detects the peripheral echo areas ar3 and ar4 in the inspection image 111A generated from the ultrasonic testing image 111, and carries out, with respect to an area sandwiched between these areas, determination of presence or absence of a defect. Thus, in accordance with the result of the determination made by the determining section 102B, the thickness calculating section 107 can calculate the distance Xi from the right edge of the peripheral echo area ar3 to the left edge of the peripheral echo area ar4 detected by the determining section 102B. Further, if a reduced scale of the inspection image 111A is obtained in advance, the thickness calculating section 107 can use the reduced scale to calculate the thickness X of the tube-to-tubesheet weld.
  • As discussed above, the determining section 102B detects the peripheral echo areas ar3 and ar4 in the process of determining presence or absence of a defect. Thus, the thickness calculating section 107 can use the result of detection of the peripheral echo areas ar3 and ar4 carried out by the determining section 102B to calculate the thickness of the inspection target portion.
  • Example of Output of Inspection Result
  • The determination result of presence or absence of a defect in the inspection target made by the information processing device 1 is output via the output section 13. Here, an example of output of an inspection result will be explained with reference to FIG. 11 . FIG. 11 shows an example of output of an inspection result.
  • The upper left part of FIG. 11 shows a defect map 300. The defect map 300 includes a doughnut-shaped area 301 showing a tube-to-tubesheet weld viewed from the tube end side, with line segments 302 indicating detected defects drawn. The defect map 300 can facilitate recognition of the distribution of the defects in the tube-to-tubesheet weld.
  • The upper right part of FIG. 11 shows a tubesheet map 400. The tubesheet map 400 schematically shows a state of a heat exchanger including a tubesheet to which many tubes are welded as shown in FIG. 2 , viewed from the tube end side. The tubesheet map 400 shows inspection results by drawing, at the location of each tube, a graphic indicating a result of a defect inspection carried out at a tube-to-tubesheet weld of that tube.
  • Specifically, a white circle is drawn at the position of a tube where no defect was detected as a result of the inspection, whereas a black circle is drawn at the position of a tube where a flaw (defect) was detected as a result of the inspection. This can facilitate recognition of the distribution of the tube-to-tubesheet welds where defects have occurred. In the tubesheet map 400, a triangle is drawn at the position of a tube which has not been inspected yet, and a square is drawn at the position of a tube which is not a target to be inspected. In this manner, various information regarding inspections may also be included in the tubesheet map 400.
  • The lower part of FIG. 11 shows an ultrasonic testing image set 500. The ultrasonic testing image set 500 includes three ultrasonic testing images (501 to 503). The ultrasonic testing image 501 is obtained by sector scanning on the tube end side, the ultrasonic testing image 502 is obtained by linear scanning, and the ultrasonic testing image 503 is obtained by sector scanning on the far side of the tube.
  • Note that the linear scanning is scanning carried out in a flaw detecting direction perpendicular to a center axis of the tube. The above-described ultrasonic testing image 111 is also obtained by linear scanning. The sector scanning on the tube end side is scanning with which an ultrasonic wave is propagated in a flaw detecting direction that is inclined toward the far side of the tube from the direction perpendicular to the center axis of the tube. The sector scanning on the far side of the tube is scanning with which an ultrasonic wave is propagated in a flaw detecting direction that is inclined toward the tube end side from the direction perpendicular to the center axis of the tube.
  • In each of these ultrasonic testing images, a reflected echo corresponding to the detected defect is marked. Indicating the ultrasonic testing images with the markings as inspection results in this manner can facilitate recognition of the position and/or the like of the defects.
  • All of the ultrasonic testing images 501 to 503 are obtained by scanning the same position in the tube-to-tubesheet weld. However, since the flaw detecting directions of the ultrasonic testing images 501 to 503 differ from each other, the defects therein look differently. Therefore, the information processing device 1 determines presence or absence of a defect in a plurality of ultrasonic testing images 111 obtained by scanning in different flaw detecting directions. Then, if the information processing device 1 determines that a defect is present in any of the flaw detecting directions, the information processing device 1 may determine, as a final determination result, that a defect is present even when no defect is found in the other flaw detecting directions. This can reduce the probability that a defect is missed. Alternatively, the information processing device 1 may determine presence or absence of a defect, with respect to a composite image that is a composite of (i) an ultrasonic testing image obtained by linear scanning and (ii) an ultrasonic testing image obtained by sector scanning.
  • The information processing device 1 may output, as an inspection result, all of or only apart of the defect map 300, the tubesheet map 400, and the ultrasonic testing image set 500. The information processing device 1 may also output, as an inspection result, information indicating the determination result regarding the type of the defect, for example. Needless to say, these are merely examples. The information processing device 1 may output a determination result in any form with which a person can recognize the content thereof.
  • Flow of Process Before Inspection
  • Before carrying out a defect inspection with use of the information processing device 1, it is necessary to construct various models used in the inspection and to determine a threshold. The description here will discuss, with reference to FIG. 12 , a flow of a process of constructing various models used in an inspection and determining a threshold. FIG. 12 is a view illustrating an example of the process of constructing various models used in an inspection and determining a threshold. These processes may be carried out by the information processing device 1 or by another computer.
  • In S1, ultrasonic testing images 111 having been subjected to smoothing processing are obtained. The ultrasonic testing images 111 include ultrasonic testing images 111 obtained from inspection targets in which a defect is present and ultrasonic testing images 111 obtained from inspection targets in which a defect is absent. The ultrasonic testing images 111 obtained from the inspection targets in which a defect is present are classified in accordance with their types of defects.
  • Note that the smoothing processing refers to a process of smoothing a change in pixel value between pixels adjacent to each other. The smoothing processing may be carried out either by the information processing device 1 or the ultrasonic testing device 7. The smoothing processing is not essential. However, it is preferable to carry out the smoothing processing, since carrying out the smoothing processing makes it easier to distinguish an echo caused by a defect from a noise component.
  • In S2, an extraction model is constructed. The extraction model is constructed by machine learning that uses training data in which an ultrasonic testing image 111 is associated with extraction area information as correct data. The extraction area information is information indicating an area to be extracted from the ultrasonic testing image 111, that is, information indicating an inspection target area. The extraction area information may be the one generated in the following manner. That is, for example, the ultrasonic testing image 111 is displayed by a display device, an operator inputs an area to be extracted, and the extraction area information is generated in accordance with the content of the input.
  • The extraction model can be constructed by any learning model suitable for extraction of an area from an image. For example, the extraction model may be constructed by You Only Look Once (YOLO) or the like that involves excellent extraction accuracy and excellent processing speed.
  • The area to be extracted can be any area that includes a tube-to-tubesheet weld, which is an inspection target portion. It is preferable that the area to be extracted also include at least apart of an area where an echo coming from its periphery appears. This is preferable because: if the inspection target portion has no defect, no feature point that can be machine-learned may be observed in that portion of the ultrasonic testing image 111; in such a case, it is difficult to construct an extraction model. For example, in the ultrasonic testing image 111 shown in FIG. 2 , an area(s) including parts of the echoes a1, a2, a6, and a7 may be set as the area to be extracted. With this, it is possible to construct the extraction model that can extract an area which includes the tube-to-tubesheet weld and the echoes coming from the periphery.
  • In S3, the extraction model constructed in S2 is used to generate images for learning (i.e., learning images) from the ultrasonic testing images 111 obtained in S1. In S2, if an extraction model is constructed by machine learning that uses, as correct data, an area including an area where an echo coming from the periphery appears, this extraction model is used to extract the inspection target area. As shown in FIG. 2 , an echo from the periphery of the inspection target portion has a feature that can be machine-learned. Therefore, with this configuration, it is possible to extract the inspection target portion automatically with high accuracy. Note that, since the learning images are generated with use of the same extraction model as that used for the inspection image 111A, the learning images have similar appearances to that of the inspection image 111A (for the appearance of the inspection image 111A, see FIG. 3 ).
  • In S4 and its subsequent step(s), the learning images generated in S3 are used for determination of a threshold and construction of a model related to the determining sections 102. In S4, a generative model is constructed. The generative model is constructed by machine learning that uses, as training data, the learning images generated from the ultrasonic testing images 111 of the inspection targets in which a defect is absent. As discussed above, the generative model may be an autoencoder. Alternatively, the generative model may be a model obtained by improving or modifying the autoencoder. For example, the generative model may be a variational autoencoder or the like.
  • If, in S2, the extraction model is constructed by machine learning that uses, as correct data, the area including the area where the echo coming from the periphery appears, the training data to be used to construct the generative model also includes the area where the echo coming from the periphery appears. In each of the ultrasonic testing images 111 of the inspection targets in which a defect is absent, the inspection target area does not include any echo, and thus does not have enough feature points to be machine-learned. However, by using the training data including the area where the echo coming from the periphery appears, it is possible to construct an appropriate generative model.
  • In S5, a threshold used by the determining section 102A to determine presence or absence of a defect is determined. Specifically, first, test images are input into the generative model constructed in S4 to generate a restored image. The test images are, among the learning images generated in S3, the ones that are not used to construct the generative model. The test images include test images generated from the ultrasonic testing images 111 of the inspection targets in which a defect is absent and test images generated from the ultrasonic testing images 111 of the inspection targets in which a defect is present. The test images generated from the ultrasonic testing images 111 of the inspection targets in which a defect is present are classified in accordance with their types.
  • Next, for the restored image generated in the above-described manner and the test images from which the restored image is generated, a difference between the restored image and the test images is calculated in pixels, and a variance of the difference is calculated. Then, a threshold is defined so that (i) variance values calculated for a plurality of test images generated from the ultrasonic testing images 111 of the inspection targets in which a defect is absent and (ii) variance values calculated for a plurality of test images generated from the ultrasonic testing images 111 of the inspection targets in which a defect is present can be distinguished from each other.
  • In S6, a reliability prediction model for the determining section 102A, which carries out determination with use of the generative model constructed in S4 and the threshold determined in S5, is constructed by machine learning. The machine learning uses training data in which a test image is associated with, as correct data, information indicating whether or not a result of determination made by the determining section 102A in accordance with that test image is correct. The test image may be the one generated from an ultrasonic testing image 111 for which presence or absence of a defect is known.
  • In response to inputting the inspection image 111A into the reliability prediction model created in this manner, a value is output which ranges from 0 to 1 and which indicates the probability that a result of determination carried out by the determining section 102A with use of the inspection image 111A is correct. Thus, the reliability determining section 103 may use an output value from the reliability prediction model as the reliability of the determination result given by the determination result 102A.
  • In S7, a type decision model is constricted by using, as training data, heat maps generated respectively from test images of various types of defects. As discussed with reference to FIG. 6 , the heat maps show features corresponding to the types of the defects. Thus, by carrying out machine learning with use of the heat maps as training data, it is possible to construct a type decision model.
  • The type decision model can be constructed by any learning model suitable for image classification. For example, the type decision model may be constructed by, e.g., convolutional neural network that has excellent image classification accuracy.
  • In S8, areas for type determination are set. Specifically, first, from the heat maps corresponding to the various types of defects and being generated in S7, defect areas where echoes caused by the defects are captured are detected. Then, in an image area of each heat map, an area where a defect area of a certain type is detected is identified as an area where the defect of the certain type appears. This process is carried out for each of the types that a user wants to carry out determination. With this, as in the example shown in FIG. 8 , it is possible to set, for each type of defect, an area where that type of defect appears.
  • Note that either of S7 and S8 may be omitted. If the process in S7 is omitted, the defect type determining section 106 determines the type of the defect in accordance with the areas set in S8. Meanwhile, if the process in S8 is omitted, the defect type determining section 106 determines the type of the defect with use of the type decision model constructed in S7.
  • In S9, a threshold to be used by the determining section 102B for numerical analysis is determined with use of the learning images generated in S3. For example, in a case where the determining section 102B carries out banalization processing, a threshold used for the banalization processing is determined.
  • In S10, a reliability prediction model for the determining section 102B, which carries out determination with use of the threshold determined in S9, is constructed by machine learning. The machine learning uses training data in which a test image is associated with, as correct data, information indicating whether or not a result of determination made by the determining section 102B in accordance with that test image is correct. The test image may be the one generated from an ultrasonic testing image 111 for which presence or absence of a defect is known.
  • In S11, a decision model used by the determining section 102C to determine presence or absence of a defect is constructed by machine learning. The machine learning uses training data in which a learning image generated in S3 is associated with, as correct data, information indicating presence or absence of a defect. With this, it is possible to construct a decision model that outputs, in response to inputting the inspection image 111A into the decision model, a value indicating the probability that a defect is present or a value indicating the probability that a defect is absent.
  • In S12, a reliability prediction model for the determining section 102C, which carries out determination with use of the decision model constructed in S11, is constructed by machine learning. The machine learning uses training data in which a test image is associated with, as correct data, information indicating whether or not a result of determination made by the determining section 102C in accordance with that test image is correct. The test image may be the one generated from an ultrasonic testing image 111 for which presence or absence of a defect is known.
  • As discussed above, the reliability prediction model for the determining section 102A can be constructed by machine learning that uses training data in which the test image for which the determining section 102A has carried out determination for presence or absence of a defect is associated with, as correct data, information indicating whether or not the result of the determination is correct. This is also true of the reliability prediction model for the determining section 102B and the reliability prediction model for the determining section 102C.
  • Each of the reliability prediction models described above has learned correspondence between (i) the test image for which the determining section 102 has carried out determination and (ii) the information indicating whether or not the result of the determination is correct. Thus, an output value obtained in response to inputting the inspection image 111A into the reliability prediction model indicates the degree of certainty of a determination result obtained when the determining section 102 carries out determination with use of the inspection image 111A.
  • Thus, the reliability determining section 103 can determine the reliabilities of the determination results of the determining sections 102 in accordance with output values obtained by inputting the inspection image 111A into the reliability prediction models for the determining sections 102, and accordingly can set reliabilities appropriate for the previous determination history.
  • Flow of Process in Inspection
  • The following description will discuss a flow of a process (determination method) in inspection with reference to FIG. 13 . FIG. 13 is a view illustrating an example of an inspection method involving use of the information processing device 1. In the description below, it is assumed that the storage section 11 stores therein ultrasonic testing images 111 each of which is an image of echoes coming from a tube-to-tubesheet weld and its periphery and measured by circumferentially moving the probe.
  • In S21, the inspection image generating section 101 generates the inspection image 111A. Specifically, the inspection image generating section 102 obtains one of the ultrasonic testing images 111 stored in the storage section 11 and inputs the one of the ultrasonic testing images 111 into the extraction model, and extracts, from the one of the ultrasonic testing images 111, an area indicated by a resulting output value so as to generate an inspection image 111A.
  • In S22 (determining step), the determining sections 102 determine presence or absence of a defect with use of the inspection image 111A generated in S21. To be more specific, in the determining section 102A, the inspection image obtaining section 1021 obtains the inspection image 111A generated in S21, and the restored image generating section 1022 generates a restored image 111B from the inspection image 111A with use of the generative model constructed in S4 in S12. Then, the defect presence/absence determining section 1023 calculates pixel-by-pixel differences between the inspection image 111A and the restored image 111B, and calculates a variance of the differences. Then, the defect presence/absence determining section 1023 determines presence or absence of a defect in accordance with whether or not the value of the variance is higher than the threshold determined in S5 in FIG. 12 . Note that, in a case where the removed image 111C and the removed image (restored) 111D have been generated, the defect presence/absence determining section 1023 calculates differences between these images.
  • The determining section 102B binarizes the inspection image 111A generated in S21 with use of the threshold determined in S9 in FIG. 12 , so as to generate a binarized image. Then, the determining section 102B detects peripheral echo areas ar3 and ar4 in the binarized image thus generated, and determines presence or absence of a defect in accordance with whether or not an area sandwiched between these areas includes a defect area.
  • Then, the determining section 102C inputs the inspection image 111A generated in S21 into the decision model constructed in S11 in FIG. 12 , and determines presence or absence of a defect in accordance with a resulting output value. For example, in a case of using the decision model that outputs the probability that a defect is present, the determining section 102C may determine that a defect is present if an output value from the decision model exceeds a given threshold. Such a threshold is also determined after the process in S11 in FIG. 12 .
  • In S23 (reliability determining step), the reliability determining section 103 determines the reliability of the determination result given by the determining section 102, with use of the inspection image 111A generated in S21. Specifically, the reliability determining section 103 determines the reliability of the determination result given by the determining section 102A, in accordance with an output value obtained by inputting the inspection image 111A into the reliability prediction model constructed in S6 in FIG. 12 .
  • For example, in a case where the reliability prediction model outputs a value which ranges from 0 to 1 and which indicates the probability that the determination result given by the determining section 102A is correct, the reliability determining section 103 may employ this value as the reliability as it is. Further, the reliability determining section 103 determines the reliability of the determination results of the determining sections 102B and 102C in a similar manner. In this manner, for each of the determining sections 102A to 102C, the reliability determining section 103 determines the reliability of the result of determination of presence or absence of a defect.
  • In S24 (comprehensive determination step), the comprehensive determination section 104 determines presence or absence of a defect with use of the determination results obtained in S22 and the reliabilities determined in S23. Specifically, the comprehensive determination section 104 determines presence or absence of a defect with use of numerical values obtained by summing up the values obtained by weighing, in accordance with their reliabilities, the numerical values indicating the determination results of the determining section 102A to 102C.
  • For example, each of the determination results of the determining sections 102A to 102C can be expressed by a numerical value “−1” (a defect is absent) or “1” (a defect is present). In this case, in a case where the reliabilities are obtained as numerical values ranging from 0 to 1, the determination results may be multiplied by the values of the reliabilities as they are.
  • Specifically, for example, assume that the determination result given by the determining section 102A indicates that a defect is present, the determination result given by the determining section 102B indicates that a defect is absent, and the determination result given by the determining section 102C indicates that a defect is present. Assume also that the reliabilities of the determination results of the determining sections 102A to 102C are 0.87, 0.51, and 0.95, respectively. In this case, the comprehensive determination section 104 carries out calculation in accordance with the following expression: 1×0.87+(−1)×0.51+1×0.95. Consequently, a numerical value of 1.31 is obtained.
  • Then, the comprehensive determination section 104 compares this numerical value with a given threshold. If the calculated numerical value is higher than the threshold, the comprehensive determination section 104 may determine that a defect is present. In a case where the result indicating that a defect is absent is expressed by a numerical value of “−1” and the result indicating that a defect is present is expressed by a numerical value of “1”, the threshold may be set at “0”, which is an intermediate value between these numerical values. In this case, since 1.31>0, a final determination result given by the comprehensive determination section 104 indicates that a defect is present.
  • In S25, the comprehensive determination section 104 records, in the inspection result data 112, the determination result obtained in S24. Then, in S26, a defect type determining process is carried out. Details of the defect type determining process will be described later with reference to FIGS. 14 and 15 .
  • In S27, the inspection image generating section 101 determines whether or not all the ultrasonic testing images 111 that are targets to be inspected have already been processed. If it is determined that there is an unprocessed ultrasonic testing image 111 (NO in S27), the process returns to S21, where the inspection image generating section 101 reads out the unprocessed ultrasonic testing image 111 from the storage section 11 and generates an inspection image 111A from the ultrasonic testing image 111. Meanwhile, if it is determined that there is no unprocessed ultrasonic testing image 111 (YES in S27), the process advances to S28.
  • In S28, the integrative detection section 108 integrates the defects detected by the comprehensive determination section 104. Then, the integrative detection section 108 records the integration result in the inspection result data 112. A method for integrating the defects is as discussed with reference to FIG. 9 , and therefore is not described here again. If there are no defects to be integrated, the processes in S28 and S29 will not be carried out, and the process advances to S30.
  • In S29, the defect length calculating section 109 calculates a length of the defects integrated by the integrative detection section 108. For example, the defect length calculating section 109 may calculate the length of the defect by multiplying (i) a length of a defect per ultrasonic testing image 111 by (ii) the number of defects integrated by the integrative detection section 108. Then, the defect length calculating section 109 records the calculation result in the inspection result data 112.
  • In S30, the thickness calculating section 107 calculates a wall thickness of the tube-to-tubesheet weld, and records the calculation result in the inspection result data 112. A method for calculating the wall thickness is as discussed above with reference to FIG. 10 , and therefore is not described here again. When the calculation results of the wall thicknesses for all the inspection image 111A are recorded, the process shown in FIG. 13 is ended.
  • Flow of Defect Type Determining Process: Type Decision Model Used
  • The following description will discuss, with reference to FIG. 14 , a flow of the defect type determining process carried out in S26 in FIG. 13 . FIG. 14 is a flowchart illustrating an example of the defect type determining process. In S41, the heat map generating section 105 generates a heat map with use of difference values (difference image) calculated when the determining section 102A determines presence or absence of a defect. Then, in S42, the heat map generating section 105 carries out threshold processing on the heat map generated in S41. The threshold processing is as discussed above with reference to FIG. 5 , and therefore is not described here again.
  • In S43, the defect type determining section 106 determines the type of the defect with use of the type decision model. Specifically, the defect type determining section 106 inputs, into the type decision model, the heat map having been subjected to the threshold processing in S42, and determines the type of the defect in accordance with a resulting output value. For example, in a case where the type decision model is constructed to output, for each type of defect, a numerical value indicating the likelihood that the defect corresponds to the type, the defect type determining section 106 may determine that the type of the defect is a type having a highest numerical value.
  • In S44, the defect type determining section 106 records, in the inspection result data 112, the determination result obtained in S43. Then, the defect type determining process is ended.
  • Flow of Defect Type Determining Process: In Accordance with Position of Defect Area
  • The defect type determining section 106 may carry out, instead of the defect type determining process shown in FIG. 14 , a defect type determining process shown in FIG. 15 . FIG. 15 is a flowchart illustrating an example of a defect type determining process that determines the type of a defect in accordance with the position of a defect area.
  • In S51, the defect type determining section 106 carries out the threshold processing on the difference values (difference image) calculated when the determining section 102A determines presence or absence of a defect. The threshold processing in S51 is similar to the threshold processing in S42 in FIG. 14 . In S52, the defect type determining section 106 detects a defect area in accordance with the difference values after the threshold processing. A method for detecting the defect area is as discussed with reference to FIG. 7 , and therefore is not described here again.
  • In S53, the defect type determining section 106 determines the type of the defect in accordance with the position of the defect area identified in S52. For example, the defect type determining section 106 may determine the type of the defect in accordance with which of the areas AR1 to AR4 shown in FIG. 8 the defect area detected in S52 is contained.
  • In S54, the defect type determining section 106 records, in the inspection result data 112, the determination result obtained in S53. Then, the defect type determining process is ended.
  • Note that the defect type determining section 106 may carry out both the defect type determining process shown in FIG. 14 and the defect type determining process shown in FIG. 15 . In this case, the defect type determining section 106 may record the determination results obtained in both. The defect type determining section 106 may integrate the two determination results together to make final determination of the type of the defect. In this case, the defect type determining section 106 may calculate the reliability of the determination result given by the defect type determining process shown in FIG. 14 and the reliability of the determination result given by the defect type determining process shown in FIG. 15 , and may yield a final determination result regarding the type of the defect in accordance with the reliabilities thus calculated. In this case, the reliabilities can be calculated in a similar manner to that for the reliability of the determination result given by the determining section 102.
  • Application Examples
  • The foregoing embodiment has dealt with the example in which presence or absence of a defect in a tube-to-tubesheet weld is determined in accordance with an ultrasonic testing image 111. However, the determination matter may be any matter, and the target data to be used for the determination may be any data selected in accordance with the determination matter. The determination matter and the target data are not limited to those adopted in the foregoing embodiment.
  • For example, the information processing device 1 is applicable to an inspection for determining presence or absence of a defect (which may also called “abnormal portion”) in an inspection target in radiographic testing (RT). In this case, an image related to an abnormal portion is detected from, in place of a radiograph, image data obtained with use of an electric device such as an imaging plate.
  • Also in this case, the determining section 102A can determine presence or absence of a defect with use of a generative model, and the determining section 102C can determine presence or absence of an abnormal portion with use of a decision model. Also, the determining section 102B can determine presence or absence of an abnormal portion by numerical analysis in accordance with pixel values, size, and/or the like of the image captured in the image data.
  • In ultrasonic testing or RT, it is possible to determine presence or absence of an abnormal portion with use of signal waveform data of an echo of an ultrasonic wave or a radioactive ray, in place of image data. Thus, the information processing device 1 is applicable to various kinds of nondestructive inspections that uses various data. Furthermore, the information processing device 1 is applicable to, in addition to the nondestructive inspections, detection of an object in a still image or a moving image and classification of the detected object, for example.
  • Variations
  • The foregoing embodiment has dealt with the example in which an output value obtained by inputting an inspection image into a reliability prediction model is used as a reliability. However, the present invention is not limited to this example. The reliability may be any one, provided that it is derived from data used by the determining section 102 for determination.
  • For example, in a case where the determining section 102B determines presence or absence of a defect with use of a binarized image obtained by binarizing an inspection image, the reliability prediction model for the determining section 102B may be a model that accepts a binarized image as input data. Meanwhile, in this case, if the determining section 102C determines presence or absence of a defect with use of the inspection image as it is, the reliability prediction model for the determining section 102C may be a model that accepts an inspection image as input data. Thus, the reliability prediction models for the determining sections 102 do not need to be constructed to accept completely the same input data.
  • The foregoing embodiment has dealt with the example in which the three determining sections 102 are employed. Alternatively, the number of determining sections 102 may be two or four or more. In the foregoing embodiment, the determination methods of the three determining sections 102 differ from each other. Alternatively, the determination methods of the three determining sections 102 may be the same. Determining sections 102 configured to carry out the same determination method may be configured to use different thresholds for determination and/or different training data to construct learnt models for determination.
  • An entity that carries out each process described in each of the foregoing embodiments can be changed as appropriate. For example, the processes in S21 (generation of an inspection image), S23 (calculation with use of a reliability decision model), S26 (determination of the type of a defect), S28 (integration of defects), S29 (calculation of a defect length), and S30 (calculation of a wall thickness) in the flowchart shown in FIG. 13 may be carried out by another information processing device. Similarly, a part or all of the processes to be executed by the determining sections 102A to 102C may be executed by another information processing device. In these cases, the number of another information processing device(s) may be one or two or more. As discussed above, the functions of the information processing device 1 can be realized by wide variety of system configurations. In a case where a system including a plurality of information processing devices is constructed, some of the plurality of information processing devices may be provided on cloud. That is, the functions of the information processing device 1 can also be realized by one information processing device or a plurality of information processing devices carrying out information processing online.
  • Software Implementation Example
  • Control blocks of the information processing device 1 (particularly, the sections included in the control section 10) can be realized by a logic circuit (hardware) provided in an integrated circuit (IC chip) or the like or can be alternatively realized by software.
  • In the latter case, the information processing device 1 includes a computer that executes instructions of an information processing program that is software realizing the foregoing functions. The computer, for example, includes at least one processor and a computer-readable storage medium storing the information processing program. An object of the present invention can be achieved by the processor of the computer reading and executing the information processing program stored in the storage medium. Examples of the processor encompass a central processing unit (CPU). The information processing device 1 may include, in addition to the processor such as CPU, a graphics processing unit (GPU). Use of GPU enables, e.g., high-speed computing involving use of the foregoing various models. The storage medium can be a “non-transitory tangible medium” such as a tape, a disk, a card, a semiconductor memory, a programmable logic circuit as well as a read only memory (ROM) or the like. The computer can further include a random access memory (RAM) in which the program is loaded. Further, the program may be made available to the computer via any transmission medium (such as a communication network and a broadcast wave) which enables transmission of the program. Note that an aspect of the present invention can also be implemented by the program in the form of a computer data signal embedded in a carrier wave which is embodied by electronic transmission.
  • The present invention is not limited to the embodiments, but can be altered by a skilled person in the art within the scope of the claims. The present invention also encompasses, in its technical scope, any embodiment derived by combining technical means disclosed in differing embodiments.
  • REFERENCE SIGNS LIST
      • 1: information processing device
      • 102A: determining section (generative model determining section)
      • 102B: determining section (numerical analysis/determination section)
      • 102C: determining section
      • 103: reliability determining section
      • 104: comprehensive determination section

Claims (7)

1. An information processing device comprising:
a reliability determining section configured to carry out, for each of determining sections configured to determine a given determination matter in accordance with single target data, a process of determining a reliability, which is an indicator indicating a degree of certainty of a determination result given by the determining section, in accordance with the target data; and
a comprehensive determination section configured to determine the given determination matter with use of the determination results and the reliabilities determined by the reliability determining section.
2. The information processing device as set forth in claim 1, wherein:
the reliability determining section configured to determine the reliabilities of the determination results given by the determining sections, in accordance with output values obtained by inputting the target data into reliability prediction models for the respective determining sections; and
each of the reliability prediction models is constructed by machine learning that uses training data in which (i) target data with respect to which a corresponding one of the determining sections has carried out determination for the determination matter is associated with, as correct data, (ii) information indicating whether or not a determination result given as a result of the determination is correct.
3. The information processing device as set forth in claim 1, wherein:
the target data is an image of an inspection target;
the determination matter is presence or absence of an abnormal portion in the inspection target;
each of the determining sections includes a generative model determining section configured to determine presence or absence of an abnormal portion with use of a generated image generated by inputting the image into a generative model; and
the generative model is constructed by machine learning that uses, as training data, an image of an inspection target not having an abnormal portion, the generative model being constructed to generate a new image having a similar feature to an image input into the generative model.
4. The information processing device as set forth in claim 3, wherein:
each of the determining sections includes a numerical analysis/determination section configured (i) to analyze pixel values in the target data, which is the image of the inspection target, so as to identify an inspection target portion in the target data and (ii) to determine presence or absence of an abnormal portion in accordance with pixel values in the inspection target portion thus identified.
5. The information processing device as set forth in claim 4, wherein:
the target data is an ultrasonic testing image which is an image of an echo of an ultrasonic wave propagated in the inspection target;
in the ultrasonic testing image, the numerical analysis/determination section identifies, as the inspection target portion, an area sandwiched between two peripheral echo areas in each of which an echo coming from a periphery of the inspection target portion appears repeatedly, and determines presence or absence of the abnormal portion in accordance with whether or not the inspection target portion thus identified includes an area constituted by pixel values each being not less than a threshold; and
the information processing device further comprises a thickness calculating section configured to calculate a thickness of the inspection target portion in accordance with a distance between the two peripheral echo areas.
6. A determination method that is to be executed by one or more information processing devices, comprising the steps of:
(i) carrying out, for each of determining sections each configured to determine a given determination matter in accordance with single target data, a process of determining a reliability, which is an indicator indicating a degree of certainty of a determination result given by the determining section, in accordance with the target data; and
(ii) determining the given determination matter with use of the determination results and the reliabilities determined in the step (i).
7. A non-transitory computer readable medium storing an information processing program configured to cause a computer to function as an information processing device recited in claim 1, the information processing program causing the computer to function as the reliability determining section and the comprehensive determination section.
US18/008,872 2020-06-10 2021-05-14 Information processing device, determination method, and information processing program Pending US20230298327A1 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
JP2020101125A JP2021196711A (en) 2020-06-10 2020-06-10 Information processing device, determination method, and information processing program
JP2020-101125 2020-06-10
PCT/JP2021/018459 WO2021251064A1 (en) 2020-06-10 2021-05-14 Information processing device, determination method, and information processing program

Publications (1)

Publication Number Publication Date
US20230298327A1 true US20230298327A1 (en) 2023-09-21

Family

ID=78847204

Family Applications (1)

Application Number Title Priority Date Filing Date
US18/008,872 Pending US20230298327A1 (en) 2020-06-10 2021-05-14 Information processing device, determination method, and information processing program

Country Status (4)

Country Link
US (1) US20230298327A1 (en)
JP (1) JP2021196711A (en)
CN (1) CN115803619A (en)
WO (1) WO2021251064A1 (en)

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2022033312A (en) * 2018-02-15 2022-02-28 株式会社三洋物産 Game machine
JP2022033308A (en) * 2018-02-15 2022-02-28 株式会社三洋物産 Game machine
JP2022033310A (en) * 2018-02-15 2022-02-28 株式会社三洋物産 Game machine
JP2022033311A (en) * 2018-02-15 2022-02-28 株式会社三洋物産 Game machine
JP2022033309A (en) * 2018-02-15 2022-02-28 株式会社三洋物産 Game machine
WO2023218537A1 (en) * 2022-05-10 2023-11-16 ファナック株式会社 Target region extraction device, method, and system
WO2024009442A1 (en) * 2022-07-06 2024-01-11 日本電信電話株式会社 Determination device, determination method, and determination program

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7132617B2 (en) * 2002-02-20 2006-11-07 Daimlerchrysler Corporation Method and system for assessing quality of spot welds
JP6472621B2 (en) * 2014-08-12 2019-02-20 株式会社Screenホールディングス Classifier construction method, image classification method, and image classification apparatus
JP6361387B2 (en) * 2014-09-05 2018-07-25 オムロン株式会社 Identification device and control method of identification device
WO2016185617A1 (en) * 2015-05-21 2016-11-24 オリンパス株式会社 Image processing device, image processing method, and image processing program
JP6772112B2 (en) * 2017-07-31 2020-10-21 株式会社日立製作所 Medical imaging device and medical image processing method

Also Published As

Publication number Publication date
CN115803619A (en) 2023-03-14
JP2021196711A (en) 2021-12-27
WO2021251064A1 (en) 2021-12-16

Similar Documents

Publication Publication Date Title
US20230298327A1 (en) Information processing device, determination method, and information processing program
US20230221286A1 (en) Inspection device, inspection method, and inspection program
US11010890B2 (en) Method for the non-destructive testing of the volume of a test object and testing device configured for carrying out such a method
US9297788B2 (en) Determination assist system of ultrasonic testing, determination assist method of ultrasonic testing, determination assist program of ultrasonic testing, and computer-readable storage medium for storing determination assist program of ultrasonic testing
US20220415020A1 (en) System and method for detection of anomalies in welded structures
JP2007271434A (en) Inspection apparatus, inspection method, inspection program, and inspection system
JP4104213B2 (en) Defect detection method
US20240161267A1 (en) Information processing device, determination method, and storage medium
JP2015190894A (en) Welding image determination device and welding image determination method
JP4956077B2 (en) Defect inspection apparatus and defect inspection method
JP7356010B2 (en) Surface texture inspection device and surface texture inspection method
KR102105503B1 (en) Method and apparatus for automatically evaluating weld quality
JP2005274444A (en) Ultrasonic flaw detection image processor, and processing method therefor
JP5150302B2 (en) Ultrasonic inspection data evaluation apparatus and ultrasonic inspection data evaluation method
JP7142825B1 (en) Image judgment method and image judgment system
JP2006226834A (en) Surface inspection device and surface inspection method
US20240046441A1 (en) Defect analysis device, defect analysis method, non-transitory computer-readable medium, and learning device
EP4109088A2 (en) Automated scan data quality assessment in ultrasonic testing
JP2789169B2 (en) Inspection method for weld surface defects of UO steel pipe
JPH06294748A (en) Surface flaw inspection method at welded part of uo steel pipe
Koskinen et al. AI for NDE 4.0–Recent use cases
JP2022026746A (en) Ultrasonic image evaluation device and ultrasonic image evaluation method
CN117999477A (en) Information processing device, method for setting determination area, and program for setting determination area

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION